Académique Documents
Professionnel Documents
Culture Documents
Version 5.5
Product Guide
P/N 302-002-366
REV 02
Copyright 2015-2016 EMC Corporation. All rights reserved. Published in the USA.
EMC believes the information in this publication is accurate as of its publication date. The information is subject to change
without notice.
The information in this publication is provided as is. EMC Corporation makes no representations or warranties of any kind with
respect to the information in this publication, and specifically disclaims implied warranties of merchantability or fitness for a
particular purpose. Use, copying, and distribution of any EMC software described in this publication requires an applicable
software license.
EMC, EMC, and the EMC logo are registered trademarks or trademarks of EMC Corporation in the United States and other
countries. All other trademarks used herein are the property of their respective owners.
For the most up-to-date regulatory document for your product line, go to EMC Online Support (https://support.emc.com).
EMC Corporation
Hopkinton, Massachusetts 01748-9103
1-508-435-1000 In North America 1-866-464-7381
www.EMC.com
Figures 7
Tables 9
Preface 11
Director failure................................................................................. 41
Intra-cluster IP management network failure.................................... 41
Intra-cluster Fibre Channel switch failure..........................................41
Inter-cluster WAN links..................................................................... 42
Power supply failures....................................................................... 42
SPS/UPS failures..............................................................................42
VPLEX Witness failure.......................................................................42
VPLEX management server failure.................................................... 43
Component IP addresses...............................................................................43
IP addresses cluster-1......................................................................44
IP addresses - cluster-2.................................................................... 45
Glossary 141
Index 149
1 VPLEX active-active........................................................................................................16
2 VPLEX family: Local and Metro....................................................................................... 18
3 Move data without disrupting service............................................................................ 20
4 Data mobility with VPLEX Local and Metro......................................................................20
5 High availability infrastructure example......................................................................... 21
6 Distributed data collaboration....................................................................................... 22
7 Architecture highlights...................................................................................................23
8 How VPLEX Witness is deployed.....................................................................................25
9 Management server - rear view...................................................................................... 30
10 Fibre Channel switch - rear view..................................................................................... 31
11 VS2: Single-engine cluster............................................................................................. 33
12 VS2: Dual-engine cluster............................................................................................... 34
13 VS2: Quad-engine cluster.............................................................................................. 35
14 Engine, rear view........................................................................................................... 36
15 VPLEX cluster independent power zones........................................................................37
16 Local mirrored volumes................................................................................................. 39
17 Component IP addresses in cluster-1.............................................................................44
18 Component IP addresses in cluster-2.............................................................................45
19 Claim storage using the GUI...........................................................................................49
20 Extents.......................................................................................................................... 54
21 Devices..........................................................................................................................55
22 Distributed devices........................................................................................................56
23 Virtual volumes............................................................................................................. 58
24 Local consistency groups with local visibility................................................................. 60
25 Local consistency group with global visibility.................................................................61
26 Path redundancy: different ports....................................................................................66
27 Path redundancy: different directors.............................................................................. 67
28 Recommended fabric assignments for front-end and back-end ports............................. 68
29 Path redundancy: different engines............................................................................... 68
30 Path redundancy: different sites.................................................................................... 69
31 High level VPLEX Witness architecture........................................................................... 70
32 Failure scenarios in VPLEX Metro configurations without VPLEX Witness........................ 71
33 VPLEX Witness and VPLEX cluster connection failures.................................................... 73
34 Unisphere Performance Monitoring Dashboard..............................................................77
35 Unisphere Performance Monitoring Dashboard - select information to view................... 78
36 Unisphere Performance Monitoring Dashboard - sample chart....................................... 78
37 Traditional view of storage arrays...................................................................................82
38 VPLEX technology refresh.............................................................................................. 83
39 Moving data with VPLEX.................................................................................................84
40 Mobility Central GUI.......................................................................................................84
41 Collaborate over distance with AccessAnywhere............................................................ 86
42 VPLEX Metro HA............................................................................................................. 88
43 VPLEX Metro HA (no cross-connect) cluster failure......................................................... 89
44 VPLEX Metro HA (no cross-connect) inter-cluster link failure...........................................89
45 VPLEX Metro HA with cross-connect............................................................................... 90
46 VPLEX Metro HA with cross-connect - host failure...........................................................91
47 VPLEX Metro HA with cross-connect - cluster failure....................................................... 91
48 VPLEX Metro HA with cross-connect - storage array failure..............................................92
49 VPLEX Metro HA with cross-connect - VPLEX Witness failure........................................... 92
50 VPLEX Metro HA with cross-connect - inter-cluster link failure........................................ 93
51 RecoverPoint architecture.............................................................................................. 94
52 RecoverPoint configurations.......................................................................................... 96
53 VPLEX Local and RecoverPoint....................................................................................... 98
54 VPLEX Local and RecoverPoint Remote - remote site is independent VPLEX cluster.........98
55 VPLEX Local and RecoverPoint remote - remote site is array-based splitter..................... 99
56 VPLEX Metro and RecoverPoint local replication.............................................................99
57 VPLEX Metro and RecoverPoint local and remote replication- local site is located at one
cluster of the VPLEX..................................................................................................... 100
58 VPLEX Metro and RecoverPoint local and remote replication- remote site is array-based
splitter.........................................................................................................................100
59 Shared VPLEX splitter.................................................................................................. 101
60 Shared RecoverPoint RPA cluster................................................................................. 101
61 Replication with VPLEX Local and CLARiiON................................................................. 102
62 Replication with VPLEX Metro and CLARiiON................................................................ 102
63 Support for Site Recovery Manager.............................................................................. 103
64 MetroPoint 2-site topology...........................................................................................104
65 MetroPoint three-site topology.....................................................................................105
66 RecoverPoint Local Replication for VPLEX Distributed Volumes at Site A and Site B and
RecoverPoint Remote Replication at Site C................................................................... 106
67 Bi-directional replication for volumes in different consistency groups and local volumes
....................................................................................................................................107
68 Active Source Site A Failure..........................................................................................108
69 Remote Site C Failure...................................................................................................109
70 MetroPoint four-site configuration............................................................................... 111
71 MetroPoint four-site configuration for a single consistency group................................ 112
72 Site A and Site B failure showing failover to Site C....................................................... 113
73 During failover, Site C and Site D become the production sites.................................... 113
74 MetroPoint - two local copies.......................................................................................116
75 Source failover............................................................................................................ 117
76 Production failover to the remote copy.........................................................................118
77 Recover production......................................................................................................119
78 VS1 single-engine cluster............................................................................................ 122
79 VS1 dual-engine cluster...............................................................................................123
80 VS1 quad-engine cluster..............................................................................................124
81 VS1 engine.................................................................................................................. 124
82 IP addresses in cluster-1..............................................................................................126
83 IP addresses in cluster-2 (VPLEX Metro)....................................................................... 127
84 Ethernet cabling - VS1 single-engine cluster................................................................ 128
85 Serial cabling - VS1 single-engine cluster.................................................................... 129
86 Fibre Channel cabling - VS1 single-engine cluster........................................................ 129
87 AC power cabling - VS1 single-engine cluster............................................................... 130
88 Ethernet cabling - VS1 dual engine cluster................................................................... 131
89 Serial cabling - VS1 dual-engine cluster....................................................................... 132
90 Fibre Channel cabling - VS1 dual-engine cluster...........................................................133
91 AC power cabling - VS1 dual-engine cluster................................................................. 134
92 Ethernet cabling - VS1 quad-engine cluster..................................................................135
93 Serial cabling - VS1 quad-engine cluster......................................................................136
94 Fibre Channel cabling - VS1 quad-engine cluster..........................................................137
95 AC power cabling - VS1 quad-engine cluster................................................................ 138
96 Fibre Channel WAN COM connections - VS1................................................................. 139
97 IP WAN COM connections - VS1................................................................................... 139
1 Typographical conventions............................................................................................ 12
2 VPLEX features and benefits.......................................................................................... 23
3 Hardware components...................................................................................................31
4 GeoSynchrony AccessAnywhere features....................................................................... 48
5 Types of data mobility operations.................................................................................. 85
6 How VPLEX Metro HA recovers from failure.....................................................................93
7 Cabling within a VPLEX cabinet.................................................................................... 127
As part of an effort to improve its product lines, EMC periodically releases revisions of its
software and hardware. Therefore, some functions described in this document might not
be supported by all versions of the software or hardware currently in use. The product
release notes provide the most up-to-date information on product features.
Contact your EMC technical support professional if a product does not function properly
or does not function as described in this document.
Note
This document was accurate at publication time. Go to EMC Online Support (https://
support.emc.com) to ensure that you are using the latest version of this document.
Purpose
This document is part of the VPLEX documentation set, and describes the VPLEX Restful
API.
Audience
This guide is intended for use by customers and service providers who use the Restful API
to configure and manage a storage environment.
Related documents (available on EMC Online Support) include:
l EMC VPLEX Release Notes for GeoSynchrony Releases
l EMC VPLEX Product Guide
l EMC VPLEX Site Preparation Guide
l EMC VPLEX Configuration Worksheet
l EMC VPLEX Configuration Guide
l EMC VPLEX Security Configuration Guide
l EMC VPLEX CLI Reference Guide
l EMC VPLEX Administration Guide
l Unisphere for VPLEX Help
l EMC VPLEX Element Manager API Guide
l EMC VPLEX Open-Source Licenses
l EMC VPLEX GPL3 Open-Source Licenses
l Procedures provided through the SolVe Desktop Generator
l EMC Host Connectivity Guides
l EMC VPLEX Hardware Installation Guide
l Various best practices technical notes available on EMC Online Support
Special notice conventions used in this document
EMC uses the following conventions for special notices:
DANGER
Indicates a hazardous situation which, if not avoided, will result in death or serious
injury.
WARNING
Indicates a hazardous situation which, if not avoided, could result in death or serious
injury.
CAUTION
Indicates a hazardous situation which, if not avoided, could result in minor or moderate
injury.
NOTICE
Note
Typographical conventions
EMC uses the following type style conventions in this document:
Technical support Go to EMC Online Support and click Service Center. You will see
several options for contacting EMC Technical Support. Note that to open a service
request, you must have a valid support agreement. Contact your EMC sales
representative for details about obtaining a valid support agreement or with questions
about your account.
Online communities Visit EMC Community Network at https://community.EMC.com for
peer contacts, conversations, and content on product support and solutions. Interactively
engage online with customers, partners, and certified professionals for all EMC products.
Your comments
Your suggestions will help to improve the accuracy, organization, and overall quality of
the user publications. Send your opinions of this document to:
DPAD.Doc.Feedback@emc.com
l VPLEX overview..................................................................................................... 16
l VPLEX product family.............................................................................................17
l Mobility.................................................................................................................19
l Availability............................................................................................................ 21
l Collaboration........................................................................................................ 21
l Architecture highlights.......................................................................................... 22
l Features and benefits............................................................................................23
l VPLEX Witness...................................................................................................... 24
l Non-disruptive upgrade (NDU)...............................................................................26
l New features in this release.................................................................................. 27
Introducing VPLEX 15
Introducing VPLEX
VPLEX overview
EMC VPLEX federates data that is located on heterogeneous storage arrays to create
dynamic, distributed and highly available data centers.
Use VPLEX to:
l Move data nondisruptively between EMC and other third party storage arrays without
any downtime for the host.
VPLEX moves data transparently and the virtual volumes retain the same identities
and the same access points to the host. There is no need to reconfigure the host.
l Collaborate over distance.
AccessAnywhere provides cache-consistent active-active access to data across VPLEX
clusters. Multiple users at different sites can work on the same data while
maintaining consistency of the dataset.
l Protect data in the event of disasters or failure of components in your data centers.
With VPLEX, you can withstand failures of storage arrays, cluster components, an
entire site failure, or loss of communication between sites (when two clusters are
deployed) and still keep applications and data online and available.
With VPLEX, you can transform the delivery of IT to a flexible, efficient, reliable, and
resilient service.
Figure 1 VPLEX active-active
VPLX-000389
VPLEX Local
VPLEX Local consists of a single cluster. VPLEX Local:
l Federates EMC and non-EMC storage arrays.
Federation allows transparent data mobility between arrays for simple, fast data
movement and technology refreshes.
l Standardizes LUN presentation and management using simple tools to provision and
allocate virtualized storage devices.
l Improves storage utilization using pooling and capacity aggregation across multiple
arrays.
l Increases protection and high availability for critical applications.
Mirrors storage across mixed platforms without host resources.
Leverage your existing storage resources to deliver increased protection and
availability for critical applications.
Deploy VPLEX Local within a single data center.
VPLEX Metro
VPLEX Metro consists of two VPLEX clusters connected by inter-cluster links with not more
than 5ms Round Trip Time (RTT). VPLEX Metro:
l Transparently relocates data and applications over distance, protects your data
center against disaster, and enables efficient collaboration between sites.
Manage all of your storage in both data centers from one management interface.
l Mirrors your data to a second site, with full access at near local speeds.
Deploy VPLEX Metro within a data center for:
l Additional virtual storage capabilities beyond that of a VPLEX Local.
l Higher availability.
Metro clusters can be placed up to 100 km apart, allowing them to be located at
opposite ends of an equipment room, on different floors, or in different fire
suppression zones; all of which might be the difference between riding through a
local fault or fire without an outage.
Deploy VPLEX Metro between data centers for:
l Mobility: Redistribute application workloads between the two data centers.
l Availability: Applications must keep running in the presence of data center failures.
l Collaboration: Applications in one data center need to access data in the other data
center.
l Distribution: One data center lacks space, power, or cooling.
Combine VPLEX Metro virtual storage and virtual servers to:
l Transparently move virtual machines and storage across synchronous distances.
l Improve utilization and availability across heterogeneous arrays and multiple sites.
Distance between clusters is limited by physical distance, by host and by application
requirements.
Mobility
VPLEX mobility allows you to move data located on either EMC or non-EMC storage arrays
simply and without disruption. Use VPLEX to simplify the management of your data center
and eliminate outages to migrate data or refresh technology.
Combine VPLEX with server virtualization to transparently move and relocate virtual
machines and their corresponding applications and data without downtime.
Relocate, share, and balance resources between sites, within a campus or between data
centers. In VPLEX Metro configurations, operations are synchronous and clusters can be
separated by up to 5 ms round trip time (RTT) latency.
Figure 4 Data mobility with VPLEX Local and Metro
MOBILITY
OBILI ACCESS ANYWHERE
Move data without impacting users Move data, VMs, and applications over distance
Heterogeneous nondisruptive Active-Active data centers avoid disasters,
technology refreshes simplify migrations and workload balancing
Use the storage and compute resources available at either of the VPLEX cluster locations
to automatically balance loads.
Move data between sites, over distance, while the data remains online and available
during the move. No outage or downtime is required.
VPLEX federates both EMC and non-EMC arrays, so even if you have a mixed storage
environment, VPLEX provides an easy solution.
Extent migrations move data between extents in the same cluster. Use extent migrations
to:
l Move extents from a hot storage volume shared by other busy extents.
l Defragment a storage volume to create more contiguous free space.
Note
Up to 25 local and 25 distributed migrations can be in progress at the same time. Any
migrations beyond those limits are queued until an ongoing migration completes.
Availability
VPLEX redundancy provides reduced Recovery Time Objective (RTO) and Recovery Point
Objective (RPO).
VPLEX features allow the highest possible resiliency in the event of an outage. The
following figure shows a VPLEX Metro configuration where storage has become
unavailable at one of the cluster sites.
Figure 5 High availability infrastructure example
Cluster A Cluster B
ACCESS ANYWHERE
X
Maintain availability and non-stop access by mirroring across locations.
Eliminate storage operatios nfrom failover.
Collaboration
Collaboration increases utilization of passive data recovery assets and provides
simultaneous access to data.
Availability 21
Introducing VPLEX
ACCESS ANYWHERE
VPLEX AccessAnywhere enables multiple users at different sites to work on the same data
while maintaining consistency of the dataset.
Traditional solutions support collaboration across distance by shuttling entire files
between locations using FTP. This is slow and contributes to network congestion for large
files (or even small files that move regularly). One site may sit idle waiting to receive the
latest data. Independent work results in inconsistent data that must be synchronized, a
task that becomes more difficult and time-consuming as your data sets increase in size.
AccessAnywhere supports co-development that requires collaborative workflows such as
engineering, graphic arts, video, educational programs, design, and research.
VPLEX provides a scalable solution for collaboration.
Architecture highlights
A VPLEX cluster consists of:
l One, two, or four VPLEX engines.
Each engine contains two directors.
Dual-engine or quad-engine clusters contain:
n One pair of Fibre Channel switches for communication between directors.
n Two Uninterruptible Power Sources (UPS) for battery power backup of the Fibre
Channel switches and the management server.
l A management server.
The management server has a public Ethernet port, which provides cluster
management services when connected to the customer network.
Brocade,
Cisco
VPLEX
Brocade,
Cisco
VPLEX conforms to established world wide naming (WWN) guidelines that can be used for
zoning.
VPLEX supports EMC storage and arrays from other storage vendors, such as HDS, HP,
and IBM.
VPLEX provides storage federation for operating systems and applications that support
clustered file systems, including both physical and virtual server environments with
VMware ESX and Microsoft Hyper-V.
VPLEX supports network fabrics from Brocade and Cisco.
Refer to the EMC Simple Support Matrix, EMC VPLEX and GeoSynchrony, available at http://
elabnavigator.EMC.com under the Simple Support Matrix tab.
Features Benefits
Mobility l Migration: Move data and applications without impact on users.
l Virtual Storage federation: Achieve transparent mobility and access within
a data center and between data centers.
l Scale-out cluster architecture: Start small and grow larger with predictable
service levels.
Availability l Resiliency: Mirror across arrays within a single data center or between data
centers without host impact. This increases availability for critical
applications.
l Distributed cache coherency: Automate sharing, balancing, and failover of
I/O across the cluster and between clusters whenever possible.
l Advanced data caching: Improve I/O performance and reduce storage array
contention.
Collaboration Distributed cache coherency: Automate sharing, balancing, and failover of I/O
across the cluster and between clusters whenever possible.
VPLEX Witness
VPLEX Witness helps multi-cluster VPLEX configurations automate the response to cluster
failures and inter-cluster link outages.
VPLEX Witness is an optional component installed as a virtual machine on a customer
host.
CAUTION
The customer host must be deployed in a separate failure domain from either of the
VPLEX clusters to eliminate the possibility of a single fault affecting both a cluster and
VPLEX Witness.
VPLEX Witness connects to both VPLEX clusters over the management IP network as
illustrated in the following figure:
VPLEX Witness observes the state of the clusters, and thus can distinguish between an
outage of the inter-cluster link and a cluster failure. VPLEX Witness uses this information
to guide the clusters to either resume or suspend I/O.
In VPLEX Metro configurations, VPLEX Witness provides seamless zero RTO fail-over for
synchronous consistency groups.
Note
VPLEX Witness works in conjunction with consistency groups. VPLEX Witness guidance
does not apply to local volumes and distributed volumes that are not members of a
consistency group.
Note
VPLEX Witness has no effect on failure handling for distributed volumes that are outside
of consistency groups or for volumes that are in asynchronous consistency groups.
Witness also has no effect on distributed volumes in synchronous consistency groups
when the preference rule is set to no-automatic-winner.
See High Availability with VPLEX Witness on page 69 for more information on VPLEX
Witness including the differences in how VPLEX Witness handles failures and recovery.
Software upgrades
VPLEX is fully redundant for:
l Ports
l Paths
l Directors
l Engines
This redundancy allows GeoSynchrony on VPLEX Local and Metro to be upgraded without
interrupting host access to storage, it does not require service window or application
disruption.
Note
You must upgrade the VPLEX management server software before upgrading
GeoSynchrony. Management server upgrades are non-disruptive.
Virtual Volume Monitor stats are row based. The CLI output of Virtual Volume Monitor
will be based on virtual volume name. The output of each monitor is a map with stats
name as the key.
l Quick vault and restore on power failure
VPLEX now reduces the amount of data at risk on power failure by starting a vault
immediately on power loss detection. It also reduces data unavailability by restarting
and allowing the VPLEX service on a director to run after the power has been restored
on it.
l Support for LDAPS server certificates
For LDAPS directories, you can now configure LDAPS server certificates for
authentication. These certificates are the files that come from the LDAPS server and
they are stored on VPLEX. These certificates create a trust mechanism between the
two systems.
l Ability to use a security certificate from a certification authority
Starting this release, you can use a security certificate, signed by a certification
authority (CA), with VPLEX. The certificates are applied for both VPN and the web
communication. A common certificate is maintained across the cluster.
l Storage volume latency monitoring
VPLEX monitors the latency for reads and writes that complete to a storage. VPLEX
samples the latency every five minutes by default and it is measured against a set
performance policy for a cluster. Storage volumes are marked degraded if the set
thresholds are exceeded. If auto-mirror isolation is enabled, VPLEX uses it to trigger
isolation on poorly performing RAID-1 mirrors.
l Resynchronization on a leg of RAID 1
The new device repair command allows you to identify and force a resynchronization
on a leg of a simple RAID 1 or a Distributed RAID 1 where one leg has been modified
on the BE storage, not through VPLEX.
l Deprecation of the ls command through the VPLEX REST API
Support for executing the ls command through the VPLEX REST API has been
deprecated. The equivalent functionality for retrieving context contents and attributes
has been provided through the VPLEX GET context URI method, which returns the
data in a structured format that is more efficiently parsed by clients of the REST API.
Ensure that you convert any existing usage of the ls command through the RESTful
API to use the VPLEX GET context URI method to avoid incompatibility. For more
information, refer to the EMC VPLEX Element Manager API Guide.
l VPLEX Cluster Witness support for VMware ESXi 6.0
You can now deploy VPLEX Cluster Witness Server on VMware ESXi 6.0.
l Enhancement to provisioning using VIAS
VPLEX release 5.5 has these enhancements for VIAS:
n VIAS now supports provisioning on VMAX3 and XtremeIO storage arrays. The
supported versions of XtremeIO are 4.0.1, 3.0.3, and 3.0.2.
n VIAS now supports Service Level Objective (SLO) and workload selection on EMC
VMAX3 storage arrays.
n Volumes and storage groups are exposed for provisioning in VIAS.
n If a provisioning job fails in VIAS, the new rollback feature brings the system to the
previous state.
n You can now specify storage groups for EMC VMAX2 arrays.
This chapter provides a high-level overview of the major hardware components in a VS2
VPLEX and how hardware failures are managed to support uninterrupted service.
VS2 Hardware 29
VS2 Hardware
Note
VS1 and VS2 hardware cannot co-exist in a cluster, except in a VPLEX Local cluster during
a non disruptive hardware upgrade from VS1 to VS2.
Management server
Each VPLEX cluster has one management server.
Figure 9 Management server - rear view
You can manage both clusters in a VPLEX Metro configuration from a single management
server.
The management server:
Each Fibre Channel switch is powered by a UPS, and has redundant I/O ports for intra-
cluster communication.
The Fibre Channel switches do not connect to the front-end hosts or back-end storage.
1, 2, or 4 VPLEX engines
A VPLEX cluster can have 1 (single), 2 (dual), or 4 (quad) engines.
l Figure 11 on page 33 shows a single-engine cluster for VS2.
l Figure 12 on page 34 shows a dual-engine cluster for VS2.
l Figure 13 on page 35 shows a quad-engine cluster for VS2.
Note
The placement of components shown for single-engine and dual-engine clusters allows
for non disruptive addition of engines to scale the cluster to a larger configuration.
The following table describes the major components of a VPLEX cluster and their
functions.
Component Description
Engine Contains two directors, with each director providing front-end and back-
end I/O connections.
Director Contains:
l Five I/O modules (IOMs), as identified in Figure 14 on page 36
l Management module for intra-cluster communication
l Two redundant 400 W power supplies with built-in fans
l CPU
l Solid-state disk (SSD) that contains the GeoSynchrony operating
environment
l RAM
Fibre Channel COM Provides intra-cluster communication support among the directors. This
switches (Dual-engine is separate from the storage I/O.
or quad-engine cluster
only)
1, 2, or 4 VPLEX engines 31
VS2 Hardware
Component Description
Power subsystem Power distribution panels (PDPs) connect to the AC power source of the
site and transfer power to the VPLEX components through power
distribution units (PDUs). This provides a centralized power interface
and distribution control for the power input lines. The PDPs contain
manual on/off power switches for their power receptacles.
Standby Power Supply One SPS assembly (two SPS modules) provides backup power to each
(SPS) engine in the event of an AC power interruption. Each SPS module
maintains power for two five-minute periods of AC loss while the engine
shuts down.
Uninterruptible Power One UPS provides battery backup for Fibre Channel switch A and the
Supply (UPS) (Dual- management server, and a second UPS provides battery backup for
engine or quad-engine Fibre Channel switch B. Each UPS module maintains power for two five-
cluster only) minute periods of AC loss while the engine shuts down.
ON ON
I I
O O
OFF OFF
ON ON
I I
O O
OFF OFF
ON ON
I I
O O
OFF OFF
Laptop tray
OFF OFF
O O
Management server
I I
ON ON
OFF OFF
O O
I I
ON ON
OFF OFF
O O
I I
ON ON
SPS 1B SPS 1A
VPLX-000226
ON ON
I I
O O
OFF OFF
ON ON
I I
O O
OFF OFF
ON ON
I I
O O
OFF OFF
Laptop tray
Fibre Channel switch B
UPS B
Fibre Channel switch A
UPS A
OFF OFF
O O
I I
Management server
ON ON
SPS 2B SPS 2A
OFF OFF
O O
I I
ON ON
SPS 1B SPS 1A
VPLX-000227
Engine 4, Director B ON
I
O
OFF
ON
I
O
OFF
Engine 4, Director A
SPS 4B SPS 4A
ON ON
I I
O O
OFF OFF
SPS 3B SPS 3A
Laptop tray
Fibre Channel switch B
UPS B
Fibre Channel switch A
UPS A
OFF OFF
O O
I I
Management server
ON ON
SPS 2B SPS 2A
OFF OFF
O O
I I
ON ON
SPS 1B SPS 1A
VPLX-000228
Management module B
Management module A
IOM B3 - Local COM
IOM A4 - reserved
Director B Director A
Depending on the cluster topology, slots A2 and B2 contain one of the following
I/O modules (IOMs) (both IOMs must be the same type):
8 Gb/s 10 Gb/s Filler module
Fibre Channel Ethernet (VPLEX Local only)
VPLX-000229
The GeoSynchrony operating system runs on the VPLEX directors, and supports:
l I/O request processing
l Distributed cache management
l Virtual-to-physical translations
l Interaction with storage arrays
WAN connectivity
For a VPLEX Metro configuration, dual inter-cluster WAN links connect the two clusters.
Clusters in a VPLEX Metro can be connected by either Fibre Channel (8 Gbps) or Gigabit
Ethernet (10 GbE).
For Fibre Channel connections, IOMs A2 and B2 contain four Fibre Channel ports.
For IP connections, IOMs A2 and B2 contain two Ethernet ports.
CAUTION
The inter cluster link carries unencrypted user data. To protect the security of the data,
secure connections are required between clusters.
AC power connection
Connect your VPLEX cluster to two independent power zones to assure a highly available
power distribution configuration.
The following figure shows AC power supplied from independent power distribution units
(PDUs):
Figure 15 VPLEX cluster independent power zones
Circuit Customer
breaker PDU 1 Circuit Customer
off (0) breaker PDU 2
Circuit breakers - Numbers
27
off (0)
Circuit breakers - Numbers
28
8
29 9
30 10
PDU 1 ...
11
CB 28 PDU 2 ...
EMC CB 9
cabinet, rear
Labels on
customer
power lines
PDPs
Power Zone
Lower B Lower A
Power Power
PDU# 1 2 PDU#
zone B zone A
Panel# Panel# (black) (gray)
CB#(s) 28 9 CB#(s)
The power supply module is field-replaceable. When power supplies are replaced one at
a time, no disruption of service occurs.
VPLEX power distribution panels (PDPs) contain manual on/off power switches for their
power receptacles.
For additional information on power requirements and practices, see the EMC Best
Practices Guide for AC Power Connections in Two-PDP Bays.
Note
All VPLEX hardware component failures are reported to the EMC Service Center to ensure
timely response and repair of these fault conditions.
I/O continues on the healthy leg of the device. When the failed/removed array is
restored, the VPLEX system uses the information in logging volumes to synchronize the
mirrors. Only the changed blocks are synchronized, therefore minimizing the inter-cluster
traffic.
The following figure shows a virtual volume mirrored between two arrays.
Figure 16 Local mirrored volumes
For maximum availability, each host can have a path to at least one front-end port on
every director.
l Use multipathing software on the host servers to ensure timely response and
continuous I/O in the presence of path failures.
l Ensure that each host has a path to each virtual volume through each fabric.
l Ensure that the fabric zoning provides hosts redundant access to the VPLEX front-end
ports.
Back end:
l Ensure that the logical unit number (LUN) mapping and masking for each storage
volume presented from a storage array to VPLEX presents the volumes out of at least
two ports from the array and on at least two different fabrics from different
controllers.
l Ensure that the LUN connects to at least two different back end ports of each director
within a VPLEX cluster.
l Active/passive arrays must have one active and one passive port zoned to each
director, and zoning must provide VPLEX with the redundant access to the array ports.
l Configure a maximum of four paths between a director and the LUN.
Note
On VS2 hardware, only 4 physical ports are available for back end connections on
each director. Refer to the VPLEX Configuration Guide for details on the hardware
configuration you are using.
Director failure
Failure of a director causes the loss of all service from that director. The second director
in the engine continues to service I/O.
VPLEX clusters containing two or more engines benefit from the additional redundancy
provided by the additional directors.
Each director within a cluster is capable of presenting the same storage.
Follow the guidelines described in Best practices: Fibre Channel ports on page 39 to
allow a host to ride through director failures by placing redundant paths to their virtual
storage through ports provided by different directors.
The combination of multipathing software on the hosts and redundant paths through
different directors of the VPLEX system allows the host to ride through the loss of a
director.
Each director is a serviceable FRU.
Note
Failure of a single subnet may result in loss of connectivity between the director and
VPLEX Witness.
Director failure 41
VS2 Hardware
Best practices
When configuring the inter-cluster network:
l Latency must be less than 5 milliseconds (ms) round trip time (RTT) for a VPLEX
Metro.
l Link speed must be a minimum of 45Mb/s of bandwidth.
The required bandwidth is dependent on the I/O pattern and must be high enough for
all writes to all distributed volumes to be exchanged between clusters.
l Switches supporting the WAN links must be configured with a battery backup UPS.
l Have physically independent WAN links for redundancy.
l Every WAN port on every director must be able to connect to a WAN port on every
director in the other cluster.
l Logically isolate the VPLEX Metro traffic from other WAN traffic using VSANs or LSANs.
l Use independent inter switch links (ISLs) for redundancy.
l Deploy VPLEX Witness in an independent failure domain.
SPS/UPS failures
Each standby power supply (SPS) is a field replaceable unit (FRU) and can be replaced
with no disruption.
SPS batteries support two sequential outages (not greater than 5 minutes) without data
loss. The recharge time for an SPS is up to 5.5 hours.
Each uninterruptible power supply (UPS) is a FRU and can be replaced with no disruption.
UPS modules support two sequential outages (not greater than 5 minutes) to the Fibre
Channel switches in a multi-engine cluster. The recharge time for a UPS to reach 90%
capacity is 6 hours.
Best practice
Best practice is to disable VPLEX Witness (while the clusters are still connected) if its
outage is expected to be long, and then revert to using preconfigured detach rules.
Once VPLEX Witness recovers, re-enable VPLEX Witness.
Refer to the EMC VPLEX CLI Guide for information about the commands to disable and
enable VPLEX Witness.
Best practice
Best practice is to disable VPLEX Witness (while the clusters are still connected) if its
outage is expected to be long, and then revert to using preconfigured detach rules.
When the management server is replaced or repaired, use the cluster-witness
enable CLI command to re-enable VPLEX Witness.
Refer to the EMC VPLEX CLI Guide for information about the commands to disable and
enable VPLEX Witness.
Component IP addresses
This section provides details about the IP addresses that are used to connect the
components within a VPLEX cluster.
1. This description applies only to synchronous consistency groups with a rule setting that identifies a specific preference.
IP addresses cluster-1
Figure 17 Component IP addresses in cluster-1
Cluster IP Seed = 1
Enclosure IDs = engine numbers
Engine 4: Engine 4:
Director 4B, A side: 128.221.252.42 Director 4A, A side: 128.221.252.41
Director 4B, B side: 128.221.253.42 Director 4A, B side: 128.221.253.41
Engine 3: Engine 3:
Director 3B, A side: 128.221.252.40 Director 3A, A side: 128.221.252.39
Director 3B, B side: 128.221.253.40 Director 3A, B side: 128.221.253.39
FC switch B 128.221.253.34
Service port
Public Ethernet port
128.221.252.2
Customer-assigned
FC switch A 128.221.252.34
Engine 2: Engine 2:
Director 2B, A side: 128.221.252.38 Director 2A, A side: 128.221.252.37
Director 2B, B side: 128.221.253.38 Director 2A, B side: 128.221.253.37
Engine 1: Engine 1:
Director 1B, A side: 128.221.252.36 Director 1A, A side: 128.221.252.35
Director 1B, B side: 128.221.253.36 Director 1A, B side: 128.221.253.35
VPLX-000242
IP addresses - cluster-2
Figure 18 Component IP addresses in cluster-2
Cluster IP Seed = 2
Enclosure IDs = engine numbers
Engine 4: Engine 4:
Director 4B, A side: 128.221.252.74 Director 4A, A side: 128.221.252.73
Director 4B, B side: 128.221.253.74 Director 4A, B side: 128.221.253.73
Engine 3: Engine 3:
Director 3B, A side: 128.221.252.72 Director 3A, A side: 128.221.252.71
Director 3B, B side: 128.221.253.72 Director 3A, B side: 128.221.253.71
FC switch B 128.221.253.66
Service port
Public Ethernet port
128.221.252.2
Customer-assigned
FC switch A 128.221.252.66
Engine 2: Engine 2:
Director 2B, A side: 128.221.252.70 Director 2A, A side: 128.221.252.69
Director 2B, B side: 128.221.253.70 Director 2A, B side: 128.221.253.69
Engine 1: Engine 1:
Director 1B, A side: 128.221.252.68 Director 1A, A side: 128.221.252.67
Director 1B, B side: 128.221.253.68 Director 1A, B side: 128.221.253.67
VPLX-000243
IP addresses - cluster-2 45
VS2 Hardware
l GeoSynchrony....................................................................................................... 48
l Management interfaces.........................................................................................49
l Provisioning with VPLEX........................................................................................ 51
l Consistency groups............................................................................................... 59
VPLEX Software 47
VPLEX Software
GeoSynchrony
GeoSynchrony is the operating system that runs on the VPLEX directors. GeoSynchrony
runs on both VS1 and VS2 hardware.
GeoSynchrony is:
l Designed for highly available, robust operation in geographically distributed
environments
l Driven by real-time I/O operations
l Intelligent about locality of access
l Designed to provide the global directory that supports AccessAnywhere
The following table summarizes features provided by GeoSynchrony and
AccessAnywhere:
Extents Storage volumes can be broken into extents and devices can be created
from these extents.
Global Visibility The presentation of a volume from one VPLEX cluster where the physical
storage for the volume is provided by a remote VPLEX cluster.
Management interfaces
In a VPLEX Metro configuration, both clusters can be managed from either management
server.
Inside VPLEX clusters, management traffic traverses a TCP/IP based private management
network.
In a VPLEX Metro configuration, management traffic traverses a VPN tunnel between the
management servers on both clusters.
Web-based GUI
VPLEXs web-based graphical user interface (GUI) provides an easy-to-use point-and-click
management interface.
The following figure shows the screen to claim storage:
Figure 19 Claim storage using the GUI
The GUI supports most of the VPLEX operations, and includes EMC Unisphere for VPLEX
online help to assist new users in learning the interface.
Management interfaces 49
VPLEX Software
VPLEX operations that are not available in the GUI, are supported by the Command Line
Interface (CLI), which supports full functionality.
VPLEX CLI
The VPLEX CLI supports all VPLEX operations.
The CLI is divided into command contexts:
l Global commands are accessible from all contexts.
l Other commands are arranged in a hierarchical context tree, and can be executed
only from the appropriate location in the context tree.
The following example shows a CLI session that performs the same tasks as shown in
Figure 19 on page 49.
Example 1 Claim storage using the CLI:
In the following example, the claimingwizard command finds unclaimed storage
volumes, claims them as thin storage, and assigns names from a CLARiiON hints file:
VPlexcli:/clusters/cluster-1/storage-elements/storage-vol
umes> claimingwizard --file /home/service/clar.txt
--thin-rebuild
Found unclaimed storage-volume
VPD83T3:6006016091c50e004f57534d0c17e011 vendor DGC:
claiming and naming clar_LUN82.
Found unclaimed storage-volume
VPD83T3:6006016091c50e005157534d0c17e011 vendor DGC:
claiming and naming clar_LUN84.
Claimed 2 storage-volumes in storage array car
Claimed 2 storage-volumes in total.VPlexcli:/clusters/cluster-1/
storage-elements/storage-vol
umes>
The EMC VPLEX CLI Guide provides a comprehensive list of VPLEX commands and detailed
instructions on using those commands.
SNMP
The VPLEX SNMP agent:
l Supports retrieval of performance related statistics as published in the VPLEX-
MIB.mib.
l Runs on the management server and fetches performance related data from
individual directors using a firmware specific interface.
l Provides SNMP MIB data for directors (local cluster only).
l Runs on Port 161 of the management server and uses the UDP protocol.
l Supports the following SNMP commands:
n SNMP Get
LDAP/AD
VPLEX administrators can choose to configure their user accounts using either:
l An external OpenLDAP or Active Directory server (which integrates with Unix using
Service for UNIX 3.5 or Identity Management for UNIX or other authentication service).
OpenLDAP and Active Directory users are authenticated by the server.
Usernames/passwords that are created on an external server are fetched from the
remote system onto the VPLEX system when they are used for the first time. They are
stored on the VPLEX system after the first use.
l The VPLEX management server
Usernames/passwords are created locally on the VPLEX system and are stored on
VPLEX.
Customers who do not want to use an external LDAP server for maintaining user accounts
can create user accounts on the VPLEX system itself.
Call-home
Call-home is a leading technology that alerts EMC support personnel of warnings in
VPLEX.
When a fault is detected in a VPLEX configuration, VPLEX automatically sends a call-home
notification to EMC and optionally to your data center support personnel.
Call-home notifications enable EMC to pro-actively engage the relevant personnel, or use
a configured ESRS gateway to resolve the problem.
If the same event occurs repeatedly on the same component, a call-home notification is
generated only for the first instance of the event, and is not generated again for the next 8
hours.
You can customize the recipients, the severity, and the text of call-home notifications to
meet your specific requirements.
Integrated storage
Integrated storage refers to storage created through the VPLEX Integrated Services
feature. This feature requires the use of Array Management Providers (AMPs) to leverage
LDAP/AD 51
VPLEX Software
Note
In this release, VIAS uses the Storage Management Initiative - Specification (SMI-S)
provider to communicate with the arrays that support integrated services to enable
provisioning.
For more information about registering AMPs and provisioning from storage pools, refer
to the provisioning chapter of the VPLEX Administration Guide.
Other storage
Other storage refers to storage from arrays that are not integrated with VPLEX through
AMPs. Because VPLEX cannot access functionality on the array, you cannot use array
functionality such as storage pools. Therefore, you can only provision from storage
volumes discovered on the array. There are two ways to provision from storage volumes:
EZ-Provisioning and advanced provisioning.
EZ-Provisioning
EZ-Provisioning uses the Provision from Storage Volumes wizard to guide you through
provisioning from selected storage volumes on the array.
Use the provisioning wizard to:
l Select or create a consistency group for the volumes
l Select the type of volume to create (distributed or local)
l Select mirroring options
l Select or create storage
l Optionally expose the virtual volumes to a host
The wizard eliminates the individual steps required to claim storage, create extents,
create devices, and then create virtual volumes on those devices. The wizard creates one
volume at a time which uses the full capacity of the selected storage volume. You can
also use this wizard to preserve existing data on a storage volume and make that data
visible to a host through a virtual volume.
Refer to Provision from storage volumes in the VPLEX Online Help for more information.
Advanced provisioning
Advanced provisioning allows you to slice (use less than the entire capacity) storage
volumes into extents. You can then use one or more of these extents to create devices,
and then virtual volumes on these devices. Advanced provisioning requires you to
perform each of these steps individually, in the order listed in the step-by-step
instructions. Use this method when you want to slice storage volumes and perform other
advanced provisioning tasks such as creating complex devices. The GUI also provides
wizards that allow you to create each of the required storage objects.
Distributed storage
Distributed storage refers to storage objects that are created by using storage from both
clusters.
To view distributed storage in the Unisphere for VPLEX GUI, hover over Provision Storage
> Distributed, and then select a distributed object to view. You can also access
distributed storage from the Distributed Storage section in the navigation pane on the
left when an individual object screen opens.
About extents
An extent is a portion of a disk.
With VPLEX, you can create an extent that uses the entire capacity of the underlying
storage volume, or just a portion of the volume.
Distributed storage 53
VPLEX Software
Figure 20 Extents
Virtual Volume
Top-level Device
Devices
Extents
Storage Volumes
About devices
Devices combine extents or other devices into a device with specific RAID techniques
such as mirroring or striping.
Figure 21 Devices
Virtual Volume
Top-level Device
Devices
Extents
Storage Volumes
Device visibility
Visibility determines which clusters know about a device.
About devices 55
VPLEX Software
A device can be visible only to the local cluster (local visibility) or to both clusters (global
visibility).
All distributed devices have global visibility.
You can create a virtual volume from extents at one cluster and make them available for
I/O (visible) at the other cluster.
A virtual volume on a top-level device that has global visibility (or is a member of a
consistency group with global visibility) can be exported in storage views on either
cluster.
Distributed devices
Distributed devices have their underlying storage arrays located at both clusters in VPLEX
Metro:
Figure 22 Distributed devices
Site A Site B
Fibre Channel
VPLX-000433
Distributed devices support virtual volumes that are presented to a host through a
storage view. From the host, the virtual volumes appear as single volumes located on a
single array.
Distributed devices are present at both clusters for simultaneous active/active read/write
access. VPLEX AccessAnywhere ensures consistency of the data between the clusters.
VPLEX distributed devices enable distributed data centers. Some of the benefits of
implementing a distributed data center include:
l Increased availability - Both data centers can serve as production workloads while
providing high availability for the other data center.
l Increased asset utilization - Passive data centers can have idle resources.
l Increased performance/locality of data access - Data need not be read from the
production site as the same data is read/write accessible at both sites.
You can configure up to 8000 distributed devices in a VPLEX system. That is, the total
number of distributed virtual volumes plus the number of top level local devices must not
exceed 8000.
Mirroring
Mirroring writes data to two or more disks simultaneously.
If one leg of a mirrored devices fails, mirroring protects data by automatically leveraging
the other disks without losing data or service.
VPLEX mirrors transparently protect applications from back-end storage array failure and
maintenance operations.
RAID-1 data is mirrored using at least two extents to duplicate the data. Read
performance is improved because either extent can be read at the same time.
VPLEX manages mirroring between heterogenous storage arrays for both local and
distributed mirroring.
Local mirroring
Local mirroring (mirroring on VPLEX Local systems) protects RAID-1 virtual volumes within
a data center.
VPLEX RAID-1 devices provide a local full-copy RAID 1 mirror of a device independent of
the host and operating system, application, and database.
Remote mirroring
Distributed mirroring (VPLEX Metro) protects distributed virtual volumes by mirroring it
between two VPLEX clusters.
Virtual Volume
Top-level Device
Devices
Extents
Storage Volumes
If needed, you can expand the capacity of a virtual volume. The expansion method
supported are as follows:
l Storage-volume
l Concatenation
Data caching
VPLEX uses advanced data caching of EMC to improve I/O performance and reduce
storage array contention. The type of data caching used for distributed volumes depends
on the VPLEX configuration. VPLEX Local and Metro configurations have round-trip
latencies of 5 ms or less, and use write-through caching.
Write-through caching
In write-through caching, a director writes to the back-end storage in both clusters before
acknowledging the write to the host.
Write-through caching maintains a real-time synchronized mirror of a virtual volume
between the two VPLEX clusters providing a recovery point objective (RPO) of zero data
loss and concurrent access to the volume through either of the clusters.
In the VPLEX user interface, write-through caching is known as synchronous cache mode.
Logging volumes
Logging volumes keep track of blocks that are written:
l During an inter-cluster link outage.
l When one leg of a distributed device becomes temporarily unreachable (is
unreachable, but not permanently removed from the SAN).
After the inter-cluster link or leg is restored, the VPLEX system uses the information in the
logging volumes to synchronize the mirrors by sending only the changed blocks across
the link.
Logging volumes also track changes during loss of a volume when that volume is one of
the mirror in a distributed device.
For a VPLEX Metro configuration, logging volumes are required at each cluster before a
distributed device can be created.
VPLEX Local configurations and systems that do not have distributed devices do not
require logging volumes.
Consistency groups
VPLEX consistency groups aggregate volumes to enable the application of a common set
of properties to the entire group.
Consistency groups aggregate up to 1,000 virtual volumes into a single entity that can be
managed as easily as an individual volume.
If all storage for an application with rollback capabilities is in a single consistency group,
the application can recover from a complete cluster failure or inter-cluster link failure with
little or no data loss. Data loss, if any, is determined by the applications data access
pattern and the consistency groups cache-mode.
All consistency groups guarantee a crash-consistent image of their member virtual
volumes. In the event of a director, cluster, or inter-cluster link failure, consistency groups
prevent possible data corruption.
Synchronous consistency groups aggregate local and distributed volumes on VPLEX Local
and VPLEX Metro systems separated by 5ms or less of latency.
l Local visibility - The local volumes in the consistency group are visible only to the
local cluster.
l Global visibility - The local volumes in the consistency group have storage at one
cluster, but are visible to both clusters.
Local visibility
Local consistency groups that have the visibility property set to the local cluster can read
and write only to their local cluster.
Figure 24 Local consistency groups with local visibility
Global visibility
Global visibility allows both clusters to receive I/O from the cluster that does not have a
local copy.
Any reads that cannot be serviced from local cache are transferred across the link. This
allows the remote cluster to have instant on demand access to the consistency group.
Detach rules
Most I/O workloads require specific sets of virtual volumes to resume on one cluster and
remain suspended on the other cluster during failures.
VPLEX includes two levels of detach rules that determine which cluster continues I/O
during an inter-cluster link failure or cluster failure.
l device-level detach rules determine which cluster continues for an individual device.
l consistency group-level detach rules determine which cluster continues for all the
member volumes of a consistency group.
There are three consistency group detach rules:
l no-automatic-winner - The consistency group does not select a winning cluster.
l active-cluster-wins - Applicable only to asynchronous consistency groups. If one
cluster was active and one was passive at the time of the outage, the consistency
group selects the active cluster as the winner.
l winner cluster-name delay seconds - Applicable only to synchronous consistency
groups. The cluster specified by cluster-name is declared the winner if an inter-cluster
link outage lasts more than the number of seconds specified by the delay.
If a consistency group has a detach-rule configured, the rule applies to all volumes in the
consistency group, and overrides any rule-sets applied to individual volumes.
In the event of connectivity loss with the remote cluster, the detach rule defined for each
consistency group identifies a preferred cluster (if there is one) that can resume I/O to the
volumes in the consistency group.
In a VPLEX Metro configuration, I/O proceeds on the preferred cluster and is suspended
on the non-preferred cluster.
Detach rules 61
VPLEX Software
WARNING
In cases where the configured winner of a detach (manual, rule, or VPLEX witness
direction) cannot present a consistent image for the consistency group at that cluster,
the detach will be prevented. A consistency group may not be able to present a
consistent image at a given cluster if one or more of its virtual volumes does not have a
healthy leg at that cluster.
This chapter describes how VPLEXs high availability and redundancy features provide
robust system integrity and resiliency.
Note
In the event of a front-end port failure or a director failure, hosts without redundant
physical connectivity to a VPLEX cluster and without multi-pathing software installed
could be susceptible to data unavailability.
Cluster
VPLEX is a true cluster architecture. That is, all devices are always available and I/O that
enters the cluster from anywhere can be serviced by any node within the cluster, while
cache and coherency is maintained for all reads and writes.
As you add more devices to the cluster, you get the added benefits of more cache,
increased processing power, and more performance.
A VPLEX cluster provides N1 fault tolerance, which means any device failure, or any
component failure can be sustained, and the cluster will continue to operate as long as
one device survives.
This is a very highly available and robust architecture, capable of sustaining even
multiple failures while the cluster still continues to provide virtualization and storage
services.
A VPLEX cluster (either VS1 or VS2) consists of redundant hardware components.
A single engine supports two directors. If one director in an engine fails, the second
director in the engine continues to service I/O. Similarly, if a VPLEX cluster contains
multiple engines, VPLEX can handle more than one failure without disrupting any services
as long as quorum (defined by set rules) is not lost.
All hardware resources (CPU cycles, I/O ports, and cache memory) are pooled.
A two-cluster configuration (Metro) offers true high availability. Operations continue and
data remains online even if an entire site fails. It also provides a high availability solution
with zero recovery point objective (RPO).
Quorum
Quorum refers to the minimum number of directors required for the cluster to service and
maintain operations.
There are different quorum rules for a cluster to become operational and start servicing
I/Os when it is booting up, also called gaining quorum. Different rules for an
operational cluster seeing director failures to either continue servicing operations and I/O
after failure handling is called maintaining quorum. Stopping servicing operations and
I/O is called losing quorum. These rules are described below:
l Gaining quorum - A non-operational VPLEX cluster gains quorum and becomes
operational when more than half of the configured directors restart and come in
contact with each other. In a single engine cluster, it refers to all the directors.
l Maintaining quorum - An operational VPLEX cluster seeing failures will continue
operating in the following scenarios:
n Director failures
If less than half of the operational directors with quorum fail.
If half of the operational directors with quorum fail, then the remaining
directors will check the operational status of the failed directors over the
management network and remain alive.
After recovering from this failure, a cluster can tolerate further similar director
failures until only one director is remaining. In a single engine cluster, a maximum
of one director failure can be tolerated.
n Intra-cluster communication failure
If there is a split in the middle, that is, half of the operational directors with
quorum lose communication with the other half of the directors, and both
halves are running, then the directors detect the operational status over the
management network and instruct half with the director with the lowest UUID
to keep running and the directors without the lowest UUID to operationally
stop.
l Quorum loss - An operational VPLEX cluster seeing failures stops operating in the
following scenarios:
n If more than half of the operational directors with quorum fail.
n If half of the operational directors with quorum fail, and the directors are unable
to determine the operation status of the other half of the directors (whose
membership includes a low UUID).
n In a dual or quad engine cluster, if all of the directors loose contact with each
other.
Quorum 65
Integrity and Resiliency
Path redundancy
Path redundancy is critical for high availability. This section describes how VPLEX delivers
resilience using multiple paths:
Different ports
Front-end ports on all directors can provide access to any virtual volume in the cluster.
Include multiple front-end ports in each storage view to protect against port failures.
When a director port fails, the host multi-pathing software seamlessly fails over to
another path through a different port, as shown in the following figure.
Figure 26 Path redundancy: different ports
Director A1 Director B1
Engine 1
Virtual Volume
VPLX-000376
Combine multi-pathing software plus redundant volume presentation for continuous data
availability in the presence of port failures.
Back-end ports, local COM ports, and WAN COM ports provide similar redundancy for
additional resilience.
Different directors
Each VPLEX engine includes redundant directors. Each director can service I/O for any
other director in the cluster due to the redundant nature of the global directory and cache
coherency.
If one director in the engine fails, then other director immediately takes over the I/O
processing from the host.
In the following figure, Director A has failed, but Director B services the host I/O that was
previously being serviced by Director A.
Director A1 Director B1
Engine 1
Virtual Volume
VPLX-000392
Note
If a director loses access to a specific storage volume, but other directors at the same
cluster have access to that volume, VPLEX can forward back-end I/O to another director
that still has access. This condition is known as asymmetric back end visibility. When
asymmetric back end visibility happens, VPLEX is considered to be in a degraded state,
and is unable provide high availability. Operations such as NDU are prevented.
Asymmetric back end visibility can also have a performance impact.
Best practices
For maximum availability:
l Present the virtual volumes through each director so that all directors except one can
fail without causing data loss or unavailability.
l Connect all directors to all storage.
l Configure paths through both an director A and a director B to ensure continuous I/O
during non-disruptive upgrade of VPLEX.
l Connect VPLEX directors to both Fibre Channel fabrics (if used) for the front-end
(host-side), and the back-end (storage array side). Isolate the fabrics.
Redundant connections from the directors to the fabrics and fabric isolation allows
VPLEX to ride through failures of an entire fabric with no disruption of service.
l Connect hosts to both fabrics and use multi-pathing software to ensure continuous
data access during failures.
l Connect I/O modules to redundant fabrics.
Different directors 67
Integrity and Resiliency
FAB-B FAB-A
VPLX-000432
Different engines
In a dual-engine or quad-engine configuration, if one engine goes down, another engine
completes the host I/O processing as shown in the following figure.
Figure 29 Path redundancy: different engines
Engine 1 Engine 2
Virtual Volume
VPLX-000393
Site distribution
When two VPLEX clusters are connected together with VPLEX Metro, VPLEX gives you
shared data access between sites. VPLEX can withstand a component failure, a site
failure, or loss of communication between sites and still keep the application and data
online and available.
VPLEX Metro ensures that if a data center goes down, or even if the link to that data
center goes down, the other site can continue processing the host I/O.
In the following figure, despite a site failure at Data Center B, I/O continues without
disruption in Data Center A.
Engine 1 Engine 1
VPLX-000394
Install the optional VPLEX Witness on a server in a separate failure domain to provide
further fault tolerance in VPLEX Metro configurations.
Note
l The customer host must be deployed in a separate failure domain from either of the
VPLEX clusters to eliminate the possibility of a single fault affecting both a cluster
and VPLEX Witness.
l The VPLEX Witness server supports round trip time latency of 1 second over the
management IP network.
In a Metro configuration, VPLEX uses rule sets to define how failures are handled. If the
clusters lose contact with one another or if one cluster fails, rule sets define which cluster
continues operation (the preferred cluster) and which suspends I/O (the non-preferred
cluster). This works for most link or cluster failures.
In the case where the preferred cluster fails, all I/O is suspended resulting in data
unavailability.
VPLEX Witness observes the state of the clusters, and thus can distinguish between a
outage of the inter-cluster link and a cluster failure. VPLEX Witness uses this information
to guide the clusters to either resume or suspend I/O.
VPLEX Witness works in conjunction with consistency groups. VPLEX Witness guidance
does not apply to local volumes and distributed volumes that are not members of a
consistency group.
In Metro systems, VPLEX Witness provides seamless zero recovery time objective (RTO)
fail-over for storage volumes in synchronous consistency groups.
Combine VPLEX Witness and VPLEX Metro to provide the following features:
Failure Domain #3
VPLEX Witness
Inter-cluster
Network A
Inter-cluster
Network B
The VPLEX Witness server must be deployed in a separate failure domain to both of the
VPLEX clusters. This deployment enables VPLEX Witness to distinguish between a site
outage and a link and to provide the correct guidance.
The VPLEX Witness software includes a client on each of the VPLEX clusters. VPLEX
Witness does not appear in the CLI until the client has been configured.
VPLX-000434
Note
Note
Failures of connections between the cluster and the VPLEX Witness VM are managed as
follows:
VPLEX VPLEX
Witness Witness
s
I/
I/
ue
nd
O
O
tin
e
Co
Su
sp
n
n
sp
Su
Co
tin
e nd
O
O
ue
I/
I/
s
s
Cluster 1 Cluster 2 Cluster 1 Cluster 2
VPLEX VPLEX
Witness Witness
s
I/
nd
I/
ue
O
O
in
Co
sp
Co
nt
nt
Su
n
Co
tin
in
O
O
ue
ue
I/
I/
s
s
VPLX-000435
When the VPLEX Witness observes a failure and provides guidance, it sticks to this
governance until both clusters report complete recovery. This is crucial in order to avoid
split-brain and data corruption.
As a result you may have a scenario where:
l Cluster-1 is isolated
l VPLEX Witness tells cluster-2 to continue I/O
l Cluster-2 becomes isolated.
Because cluster-2 has previously received guidance to proceed from VPLEX Witness, it
proceeds even when it is isolated. In the meantime, if cluster-1 reconnects with the
VPLEX Witness server, the VPLEX Witness server tells cluster-1 to suspend. In this case,
because of event timing, cluster-1 is connected to VPLEX Witness but it is suspended,
while cluster-2 is isolated but it is proceeding.
Higher availability
Combine VPLEX Witness with VMware and cross cluster connection to create even higher
availability.
ALUA
Asymmetric Logical Unit Access (ALUA) routes I/O of the LUN directed to non-active/failed
storage processor to the active storage processor without changing the ownership of the
LUN.
Each LUN has two types of paths:
l Active/optimized paths are direct paths to the storage processor that owns the LUN.
Active/optimized paths are usually the optimal path and provide higher bandwidth
than active/non-optimized paths.
l Active/non-optimized paths are indirect paths to the storage processor that does not
own the LUN through an interconnect bus.
I/Os that traverse through the active/non-optimized paths must be transferred to the
storage processor that owns the LUN. This transfer increases latency and has an
impact on the array.
VPLEX detects the different path types and performs round robin load balancing across
the active/optimized paths.
VPLEX supports all three flavors of ALUA:
l Explicit ALUA - The storage processor changes the state of paths in response to
commands (for example, the Set Target Port Groups command) from the host (the
VPLEX backend).
The storage processor must be explicitly instructed to change a paths state.
If the active/optimized path fails, VPLEX issues the instruction to transition the
active/non-optimized path to active/optimized.
There is no need to failover the LUN.
l Implicit ALUA -The storage processor can change the state of a path without any
command from the host (the VPLEX back end).
If the controller that owns the LUN fails, the array changes the state of the active/non-
optimized path to active/optimized and fails over the LUN from the failed controller.
On the next I/O, after changing the paths state, the storage processor returns a Unit
Attention Asymmetric Access State Changed to the host (the VPLEX backend).
VPLEX then re-discovers all the paths to get the updated access states.
l Implicit/explicit ALUA - Either the host or the array can initiate the access state
change.
Storage processors support implicit only, explicit only, or both.
Metadata volumes
Meta-volumes store VPLEX metadata, including virtual-to-physical mappings, data about
devices, virtual volumes, and system configuration settings.
Metadata is stored in cache and backed up on specially designated external volumes
called meta-volumes.
After the meta-volume is configured, updates to the metadata are written to both the
cache and the meta-volume when the VPLEX configuration is modified.
Each VPLEX cluster maintains its own metadata, including:
l The local configuration for the cluster.
l Distributed configuration information shared between clusters.
At system startup, VPLEX reads the metadata and loads the configuration information
onto each director.
When you make changes to the system configuration, VPLEX writes these changes to the
metadata volume.
If VPLEX loses access to the metadata volume, the VPLEX directors continue
uninterrupted, using the in-memory copy of the configuration. VPLEX blocks changes to
the system until access is restored or the automatic backup meta-volume is activated.
Meta-volumes experience high I/O only during system startup and upgrade.
I/O activity during normal operations is minimal.
Best practices
For best resilience:
l Allocate storage volumes of 78Gb for the metadata volume.
l Configure the metadata volume for each cluster with multiple back-end storage
volumes provided by different storage arrays of the same type.
l Use the data protection capabilities provided by these storage arrays, such as RAID 1
to ensure the integrity of the system's metadata.
l Create backup copies of the metadata whenever configuration changes are made to
the system.
l Perform regular backups of the metadata volumes on storage arrays that are separate
from the arrays used for the metadata volume.
Logging volumes
Logging volumes keep track of blocks written:
l During an inter-cluster link outage.
CAUTION
If no logging volume is accessible, then the entire leg is marked as out-of-date. A full re-
synchronization is required once the leg is reattached.
The logging volumes on the continuing cluster experience high I/O during:
l Network outages or cluster failures
l Incremental synchronization
When the network or cluster is restored, VPLEX reads the logging volume to determine
what writes to synchronize to the reattached volume.
There is no I/O activity during normal operations.
Best practices
On a VPLEX Metro configuration:
l Place logging volumes on the fastest storage available.
l Stripe logging volumes across several disks to accommodate the high level of I/O
that occurs during and after outages.
l Configure at least 1 Gb of logging volume space for every 16 TB of distributed device
space.
Slightly more space is required if the 16 TB of distributed storage is composed of
multiple distributed devices, because a small amount of non-logging information is
also stored for each distributed device.
Global cache
Memory systems of individual directors ensure durability of user and critical system data.
Synchronous systems (write-through cache mode) leverage the back-end array by writing
user data to the array. An acknowledgment for the written data must be received before
the write is acknowledged back to the host.
VPLEX security features help keep your system safe from intruders.
Performance monitoring
VPLEX performance monitoring provides a customized view into the performance of your
system. You decide which aspects of the system's performance to view and compare.
You can view and assess the VPLEX performance using these methods:
l Unisphere Performance Monitoring Dashboard, which shows real-time performance
monitoring data for up to one hour of history.
l Performance statistics collection using the CLI and the API. These methods let you
collect and view the statistics, and export them to an external application for
analysis.
l Monitoring with Simple Networking Management Protocol (SNMP).
You decide which aspects of the system's performance to view and compare:
Performance monitoring 77
Integrity and Resiliency
Performance information is displayed as a set of charts. For example, the following figure
shows front-end throughput for a selected director:
Figure 36 Unisphere Performance Monitoring Dashboard - sample chart
For additional information about the statistics available through the Performance
Monitoring Dashboard, see the EMC Unisphere for VPLEX online help available in the
VPLEX GUI.
Security features
The VPLEX management server operating system (OS) is based on a Novell SUSE Linux
Enterprise Server 11 distribution.
The operating system has been configured to meet EMC security standards by disabling
or removing unused services, and protecting access to network services through a
firewall.
Security features include:
l SSH Version 2 to access the management server shell
l Customizable password policies
l LDAP authentication using LDAP v3 protocol
l IPv6 support
l HTTPS to access the VPLEX GUI
l IPSec VPN inter-cluster link in a VPLEX Metro configuration
l IPSec VPN to connect each cluster of a VPLEX Metro to the VPLEX Witness server
l SCP to copy files
l Support for separate networks for all VPLEX cluster communication
l Defined user accounts and roles
l Defined port usage for cluster communication over management server
l Certificate Authority (CA) certificate (default expiration 5 years)
l Two host certificates (default expiration 2 years)
Security features 79
Integrity and Resiliency
CAUTION
The WAN-COM inter-cluster link carries unencrypted user data. To ensure privacy of
the data, establish an encrypted VPN tunnel between the two sites.
For more information about security features and configuration see the EMC VPLEX
Security Configuration Guide.
l Technology refresh................................................................................................ 82
l Mobility.................................................................................................................83
l Collaboration........................................................................................................ 86
l VPLEX Metro HA.....................................................................................................87
l Redundancy with RecoverPoint..............................................................................94
l MetroPoint.......................................................................................................... 103
Technology refresh
In typical IT environments, migrations to new storage arrays (technology refreshes)
require that the data that is being used by hosts be copied to a new volume on the new
array. The host must then be reconfigured to access the new storage. This process
requires downtime for the host.
Migrations between heterogeneous arrays can be complicated and may require
additional software or functionality. Integrating heterogeneous arrays in a single
environment is difficult and requires a staff with a diverse skill set.
The following figure shows the traditional view of storage arrays with servers attached to
the redundant front end and storage (Array 1 and Array 2) connected to a redundant
fabric at the back end.
Figure 37 Traditional view of storage arrays
Array 1
Server
No Federation
Server
Array 2
VPLX-000381
When VPLEX is inserted between the front-end and back-end redundant fabrics, VPLEX
appears as the target to hosts and as the initiator to storage.
This abstract view of storage is very helpful while replacing the physical array that is
providing storage to applications.
With VPLEX, because the data resides on virtual volumes, it can be copied
nondisruptively from one array to another without any downtime. There is no need to
reconfigure the host; the physical data relocation is performed by VPLEX transparently
and the virtual volumes retain the same identities and the same access points to the
host.
In the following figure, the virtual disk is made up of the disks of Array A and Array B. The
site administrator has determined that Array A has become obsolete and should be
replaced with a new array. Array C is the new storage array. Using Mobility Central, the
administrator:
l Adds Array C into the VPLEX cluster.
l Assigns a target extent from the new array to each extent from the old array.
l Instructs VPLEX to perform the migration.
VPLEX copies data from Array A to Array C while the host continues its access to the
virtual volume without disruption.
After the copy of Array A to Array C is complete, Array A can be decommissioned:
Virtual Volume
VPLX-000380
Because the virtual machine is addressing its data to the abstracted virtual volume, its
data continues to flow to the virtual volume without any need to change the address of
the data store.
Although this example uses virtual machines, the same is true for traditional hosts. Using
VPLEX, the administrator can move data used by an application to a different storage
array without the application or server being aware of the change.
This allows you to change the back-end storage arrays transparently, without interrupting
I/O.
VPLEX makes it easier to replace heterogeneous storage arrays on the back-end.
Mobility
Use VPLEX to move data between data centers, relocate a data center or consolidate
data, without disrupting host application access to the data.
Mobility 83
VPLEX Use Cases
MOBILITY
Cluster A Cluster B
ACCESS ANYWHERE
The source and target arrays can be in the same data center (VPLEX Local) or in different
data centers separated by up to 5ms (VPLEX Metro).
With VPLEX, source and target arrays can be heterogeneous.
When you use VPLEX to move data, the data retains its original VPLEX volume identifier
during and after the mobility operation. No change in volume identifiers eliminates
application cut over. The application continues to use the same storage, unaware that
it has moved.
There are many types and reasons to move data:
l Move data from a hot storage device.
l Move applications from one storage device to another.
l Move operating system files from one storage device to another.
l Consolidate data or database instances.
l Move database instances.
l Move storage infrastructure from one physical location to another.
With VPLEX, you no longer need to spend significant time and resources preparing to
move data and applications. You do not have to accept a forced outage and restart the
application after the move is completed.
Instead, a move can be made instantly between sites, over distance, and the data
remains online and available during the move without any outage or downtime.
Considerations before moving the data include the business impact, type of data to be
moved, site locations, total amount of data, and schedules.
To move storage:
l Display and select the extents (for extent mobility) or devices (for device mobility) to
move.
Extent Moves data from one extent to another extent (within a cluster).
Device Moves data from one device to another device (within a cluster).
Batch Moves data using a migration plan file. Create batch migrations to automate routine
tasks.
l Use batched extent migrations to migrate arrays within the same cluster where the
source and destination have the same number of LUNs and identical capacities.
l Use batched device migrations to migrate to dissimilar arrays and to migrate devices
between clusters in a VPLEX Metro.
Best practices
All components of the system (virtual machine, software, volumes) must be available and
in a running state.
Data mobility can be used for disaster avoidance, planned upgrade, or physical
movement of facilities.
Collaboration
If you require distributed data collaboration, VPLEX can provide a significant advantage.
Traditional collaboration across distance required files to be saved at one location and
then sent to another site. This is slow, incurs bandwidth costs for large files or even small
files that move regularly. Also, it negatively impacts resource productivity as sites sit idle
while they wait to receive the latest changes. Independent work quickly leads to version
control problems as multiple people working at the same time are unaware of each
others most recent changes. Merging independent work is time-consuming, costly, and
grows more complicated as the dataset gets larger.
Current applications for information sharing over distance are not suitable for
collaboration in Big Data environments. Transfer of hundreds of Gb or TB of data across
WAN is inefficient, especially if you need to modify only a small portion of a huge data
set, or use it for analysis.
VPLEX enables multiple users at different sites to work on the same data, and maintain
consistency in the dataset when changes are made.
With VPLEX, the same data can be accessed by all users at all times, even if users are at
different sites. The data is shared and not copied, so that a change made in one site
shows up immediately at the other site.
VPLEX need not ship the entire file back and forth like other solutions. It only sends the
changed updates as they are made, greatly reducing bandwidth costs and offering
significant savings over other solutions.
With VPLEXs Federated AccessAnywhere, the data remains consistent, online, and
always available.
Figure 41 Collaborate over distance with AccessAnywhere
ACCESS ANYWHERE
Deploy VPLEX to enable real collaboration between teams located at different sites.
Local consistency groups with global visibility allow the remote cluster to read and write
to the local consistency group. The remote cluster has instant on-demand access to the
consistency group.
Local consistency groups with global visibility might contain only local volumes and the
cache mode must be synchronous. Thus, latency between clusters for this configuration
cannot exceed 5ms RTT.
VPLEX Metro HA
VPLEX Metro High Availability (HA) configurations consist of a VPLEX Metro system
deployed in conjunction with VPLEX Witness. There are two types of Metro HA
configurations:
l VPLEX Metro HA can be deployed in places where the clusters are separated by 5 ms
latency RTT or less.
l VPLEX Metro HA combined with Cross Connect between the VPLEX clusters and hosts
can be deployed where the clusters are separated by 1 ms latency RTT or less.
This section describes the benefits of these two configurations.
AccessAnywhere enables both clusters to provide simultaneous coherent read/write
access to the same virtual volume. Paths are up and the storage is available during
normal operation and not only after failover.
VPLEX Metro HA 87
VPLEX Use Cases
Note
The VPLEX clusters in VPLEX Metro HA configuration must be within 5 ms RTT latency.
In this deployment, virtual machines can write to the same distributed device from either
cluster and move between two geographically disparate locations.
If you use VMware Distributed Resource Scheduler (DRS) to automate load distribution on
virtual machines across multiple ESX servers, you can move a virtual machine from an
ESX server attached to one VPLEX cluster to an ESX server attached to the second VPLEX
cluster, without losing access to the underlying storage.
Note
VPLEX Metro HA combined with cross-connect eliminates RTO for most of the failure
scenarios.
Cluster failure
If a VPLEX cluster fails:
l VPLEX Witness guides the surviving cluster to continue.
l VMware re-routes I/O to the surviving cluster.
l No disruption to I/O.
Figure 47 VPLEX Metro HA with cross-connect - cluster failure
WARNING
Although this failure causes no disruption to the clusters or the virtual machines, it
makes the configuration vulnerable to a second failure of a major component. If a cluster
or inter-cluster link failure occurs while VPLEX Witness is not available, distributed
devices are suspended at both clusters. Therefore, if VPLEX Witness will be unavailable
for an extended period, best practice is to disable the VPLEX Witness and allow the
devices to use their configured detach rules.
The following table summarizes how VPLEX HA with cross-connect manages failures.
VPLEX cluster failure VPLEX Witness detects the failure and enables all volumes on the surviving
(Site 1) cluster.
Inter-cluster link l If the cross-connects use different physical links from those used to
failure connect the VPLEX clusters, applications are unaffected. Every volume
continues to be available in one data center or the other.
l If the cross-connect links use the same physical links as those used to
connect the VPLEX clusters, an application restart is required.
Storage array failure Applications are unaffected. VPLEX dynamically redirects I/O to the
mirrored copy on the surviving array.
Note
This example assumes that all distributed volumes are also mirrored on
the local cluster. If not, then the application remains available because the
data can be fetched or sent from or to the remote cluster. However, each
read/write operation now incurs a performance cost.
Failure of Both clusters call home. As long as both clusters continue to operate and
VPLEX Witness there is no inter-cluster link partition, applications are unaffected.
CAUTION
Note
RecoverPoint integration is offered for VPLEX Local and VPLEX Metro configurations.
The VPLEX splitter works with a RecoverPoint Appliance (RPA) to orchestrate the
replication of data either remotely or locally, or both.
Figure 51 RecoverPoint architecture
The VPLEX splitter enables VPLEX volumes in a VPLEX Local or VPLEX Metro to mirror I/O
to a RecoverPoint Appliance.
RecoverPoint Appliances
The RecoverPoint Appliance manages all aspects of data protection. One RPA can
manage multiple replication sets (production volume and one or more replica volumes to
which it replicates), each with differing policies.
For redundancy, a minimum of two RPAs are installed at each site, located in the same
facility as the host and storage. Each site can have as many as eight RPAs.
The set of RPAs installed at each site is referred to as an RPA cluster. If one RPA in a
cluster fails, the functions provided by the failed RPA are automatically moved to one or
more of the remaining RPAs.
In the case of remote replication or mixed local and remote replication, RPAs are required
at both sites. The RPAs at the production site transfer the split I/O to the replica site.
The RPAs at the replica site distribute the data to the replica storage.
In the event of a failover, these roles can be reversed. The same RPA can serve as the
production RPA for one RecoverPoint consistency group and as the replica RPA for
another.
RecoverPoint Volumes
RecoverPoint requires volumes to act as repository volumes, journal volumes, and
production volumes. All RecoverPoint volumes can be hosted on VPLEX. In practice, some
volumes can be hosted on VPLEX and others can be hosted on non-VPLEX storage. For
example, the repository volume for an existing RPA cluster cannot be moved. If you are
installing VPLEX into an existing RecoverPoint configuration, the repository volume is
already configured on non-VPLEX storage.
Note
Recovery/Failover
RecoverPoint provides several image access methods on a replica journal to enable read/
write access to a selected point-in-time at a replica. There are four image access modes:
l Logged (physical) access - Used for production recovery, failover, testing, and cloning
a replica.
l Virtual (instant) access - Used for single file recovery or light testing. Used to gain
access to the replica data immediately, or when I/O performance is not important.
l Virtual (instant) access with roll - Used for production recovery, failover, or
processing with a high write-rate. Used when the point-in-time image is far from the
current point-in-time (and would take too long to access in logged access mode).
l Direct access - This access mode can only be enabled after logged access, or virtual
access with roll, are enabled. Used for extensive processing with a high write-rate,
RecoverPoint Appliances 95
VPLEX Use Cases
when image access is needed for a long period of time (and may not have the journal
space to support all of the data written to the image access log in this time), and
when it is not required to save the history in the replica journal (the replica journal is
lost after direct access).
Note
In the current release, virtual (instant) access and virtual (instant) access with roll are not
supported by the VPLEX splitter.
A bookmark is a label applied to a snapshot so that the snapshot can be explicitly called
(identified) during the recovery processes (during image access).
Bookmarks are created through the RecoverPoint GUI and can be created manually, by
the user, or automatically, by the system. Bookmarks created automatically can be
created at pre-defined intervals or in response to specific system events. Parallel
bookmarks are bookmarks that are created simultaneously across multiple consistency
groups.
RecoverPoint configurations
RecoverPoint supports three replication configurations:
Figure 52 RecoverPoint configurations
Local Replication
In a Local Replication, RecoverPoint continuously replicates data within the same site.
Every write is kept in the journal volume, allowing recovery to any point in time. By
default, snapshot granularity is set to per second, so the exact data size and contents are
determined by the number of writes made by the host application per second. If
necessary, the snapshot granularity can be set to per write.
Remote Replication
In these configurations, data is transferred between a local and a remote site over Fibre
Channel or a WAN. The RPAs, storage, and splitters are located at both the local and the
remote site.
By default, the replication mode is set to asynchronous, and the snapshot granularity is
set to dynamic, so the exact data size and contents are determined by the policies set by
the user and system performance. This provides application consistency for specific
points in time.
Synchronous replication is supported when the local and remote sites are connected
using Fibre Channel or IP.
RecoverPoint/VPLEX configurations
RecoverPoint can be configured on VPLEX Local or Metro systems as follows:
l
RecoverPoint/VPLEX configurations 97
VPLEX Use Cases
Figure 55 VPLEX Local and RecoverPoint remote - remote site is array-based splitter
VPLEX Metro and RecoverPoint with both Local and Remote Replication
In VPLEX Metro/RecoverPoint with both local and remote replication configurations, I/O
is:
l Written to both VPLEX clusters (as part of normal VPLEX operations).
l Split on one VPLEX cluster to replica volumes located both at the cluster and at a
remote site.
RPAs are deployed at one VPLEX cluster and at a third site.
The third site can be an independent VPLEX cluster:
Figure 57 VPLEX Metro and RecoverPoint local and remote replication- local site is located at one
cluster of the VPLEX.
This configuration supports unlimited points in time, with granularity up to a single write,
for local and distributed VPLEX virtual volumes.
l RecoverPoint Appliances can (and for MetroPoint must) be deployed at each VPLEX
cluster in a Metro system. For MetroPoint replication, a different RecoverPoint cluster
must be viewing each exposed leg of the VPLEX distributed volume. The two
RecoverPoint clusters become the active and standby sites for the MetroPoint group.
l All RecoverPoint protected volumes must be on the preferred cluster, as designated
by VPLEX consistency group-level detach rules.
l Customers can recover from operational disasters by quickly returning to any point-
in-time on the VPLEX cluster where the RPAs are deployed or at the third site.
l Application event aware based rollback is supported on VPLEX Metro distributed/
local virtual volumes for Microsoft SQL, Microsoft Exchange, and Oracle database
applications.
l If the VPLEX cluster fails, then the customers can recover to any point in time at the
remote replication site. Recovery at the remote site to any point in time can be
automated through integration with MSCE and VMware Site Recovery Manager (SRM).
l This configuration can simulate a disaster at the VPLEX cluster to test RecoverPoint
disaster recovery features at the remote site.
In the configuration depicted below, host writes to the distributed virtual volumes are
written to the both legs of the distributed RAID 1 volume. Additionally, a copy of the I/O is
sent to the RPA. RPA then distributes to the replica on the CLARiiON array at a remote
disaster recovery site:
Figure 62 Replication with VPLEX Metro and CLARiiON
When an outage occurs in VPLEX Local or VPLEX Metro configurations, the virtual
machines can be restarted at the replication site with automatic synchronization to the
VPLEX configuration when the outage is over.
MetroPoint
VPLEX GeoSynchrony configured with RecoverPoint in a VPLEX Metro provides the
MetroPoint topology. This MetroPoint topology provides a 3-site or 4-site solution for
continuous availability, operational and disaster recovery, and continuous data
protection. MetroPoint also supports a 2-site topology with the ability to expand to a third
remote site in future.
The MetroPoint topology provides full RecoverPoint protection of both sides of a VPLEX
distributed volume across both sides of a VPLEX Metro configuration, maintaining
replication and protection at a consistency group level, even when a link from one side of
the VPLEX Metro to the replication site is down.
In MetroPoint, VPLEX Metro and RecoverPoint replication are combined in a fully
redundant manner to provide data protection at both sides of the VPLEX Metro and at the
replication site. With this solution, data is replicated only once from the active source site
to the replication site. The standby source site is ready to pick up and continue
replication even under a complete failure of the active source site.
MetroPoint combines the high availability of the VPLEX Metro with redundant replication
and data protection of RecoverPoint. MetroPoint protection allows for one production
copy of a distributed volume on each Metro site, one local copy at each Metro site, and
one remote copy for each MetroPoint consistency group. Each production copy can have
multiple distributed volumes.
MetroPoint offers the following benefits:
l Full high availability for data access and protection.
l Continuous data protection and disaster recovery.
l Operational recovery at all three sites for redundancy.
l Efficient data transfer between VPLEX Metro sites and to the remote site.
l Load balancing across replication links and bi-directional replication.
l Out of region data protection with asynchronous replication.
l Any-Point-in-Time operational recovery in the remote site and optionally in each of
the local sites. RecoverPoint provides continuous data protection with any-point-in-
time recovery.
MetroPoint 103
VPLEX Use Cases
l Full support for all operating systems and clusters normally supported with VPLEX
Metro.
l Support for a large variety of EMC and third-party storage arrays.
Journal and repository volumes are local, not distributed. This enhances performance and
reduces load on the inter-cluster link resulting from high I/O traffic needed for journals.
The three-site MetroPoint topology features a VPLEX Metro (Site A and Site B) and a
remote Site C. Site C can be any valid array based splitter option including:
l VPLEX Local
l VPLEX Metro
l VNX
l VNX2
l VMAX
Figure 65 MetroPoint three-site topology
The above figure illustrates the basic MetroPoint topology, with a VPLEX Metro system
consisting of VPLEX sites A and B, and a remote Site C. Each site is protected by a
RecoverPoint cluster. In this example, both production copies of the VPLEX distributed
volume are protected by RecoverPoint clusters at sites A and B. The remote replicating
RecoverPoint cluster is Site C.
One RecoverPoint cluster at Metro Site A is configured as the active source of remote
replication, while the other RecoverPoint cluster at Metro Site B is the standby source.
Both active and standby source sites maintain up-to-date journals. If the active source
cluster fails, the standby source site instantly switches over and becomes the active
source of remote replication after a short initialization period.
MetroPoint also allows for local copies of the distributed volume at each site in addition
to the remote copy. The local copies are independent of one another.
For each volume protected, a single consistency group spans the three related sites.
MetroPoint consistency groups are referred to as MetroPoint groups. The MetroPoint
group has a production copy on both Metro sites. The production copy at the active
source site is the only one that replicates to the remote site.
Figure 66 RecoverPoint Local Replication for VPLEX Distributed Volumes at Site A and Site B and
RecoverPoint Remote Replication at Site C
MetroPoint configurations must adhere to both VPLEX and RecoverPoint best practices for
the distances between sites. In a VPLEX Metro, the production copies should be within
synchronous distance. For RecoverPoint, the site for the remote copy can be within
synchronous or asynchronous distance.
Distributed volumes
RecoverPoint protects VPLEX distributed volumes on both Metro Sites A and B for
redundancy, and protects copies on the remote site.
Local (non-distributed) volumes
VPLEX local (non-distributed) volumes can be replicated by RecoverPoint from any Site A,
B, or C, and are protected by RecoverPoint Local Replication.
Local volumes can be replicated from Site C to Site A or Site B if the volumes are
independent from the VPLEX volumes that are distributed from Sites A and B.
Figure 67 Bi-directional replication for volumes in different consistency groups and local volumes
Distributed RAID 1 volumes in MetroPoint groups from Site A and Site B replicate to
remote Site C.
Local volumes on Site A, B, or C can replicate to a target volume on a site other than their
local site. For example, a local volume on Site C can replicate to Site B, or a local volume
in a different VPLEX consistency group in Site A can replicate to Site C. RecoverPoint
supports bi-directional replication for volumes in different consistency groups.
Failure scenarios
MetroPoint protects applications and data from any failure scenario including:
l Any single component failure
l A complete site failure
l A regional failure that affects two sites
l Data unavailability due to user errors
l Data corruption
For disaster recovery, when creating MetroPoint groups, the best practice is to set the
group policy for the preferred source of remote replication to the default Follow VPLEX
bias rules. This provides added protection in that RecoverPoint will select the side that
the VPLEX consistency group will declare as the winner site in the event of link failures in
the VPLEX Metro system and prevent unnecessary swapping of the source of replication.
If a disaster occurs with either the RecoverPoint cluster or the VPLEX Metro which
prevents the active source site from replicating, it can select the other site for the remote
replication to continue.
If Site A failed, operations would continue uninterrupted at Site B. Replication from Site A
would switch over to replicate from Site B to remote Site C without losing replication
history. Site B would now be the active source site until Site A is restored on a
consistency group level.
When recovering from a local copy after a switchover, the remote copy is temporarily
disabled. Both local copies are initialized when one is restored, and the remote copy will
be updated after recovery and a short initialization.
Asynchronous WAN replication load balancing is done using VPLEX bias rules. By default,
the active and standby source sites are based on VPLEX bias rules that are configured at
the consistency group level. Using this method, the active source site is the site that has
bias. Using the bias rules, replication traffic can be load balanced across the links at a
consistency group level. Load balancing between the two source sites improves resource
distribution and provides better high availability.
If the RecoverPoint cluster fails at the remote site (Site C), local replication continues on
distributed RAID 1 volumes on both sides of the Metro (Site A and Site B). There is no
copy, no replication, and no image access at Site C.
When the remote RecoverPoint cluster is restored, RecoverPoint automatically
resynchronizes and replication resumes at the remote site.
The MetroPoint topology is specifically designed to provide protection in the following
scenarios:
l VPLEX fracture, inter-cluster link failure
l VPLEX site failure on the RecoverPoint active production copy
l RecoverPoint failure (active source)
l Data corruption
VPLEX fracture, inter-cluster link failure
A fracture occurs when the communication link between two VPLEX Metro sites is down.
When this occurs:
l Replication continues through the VPLEX winner site.
l The preferred RecoverPoint cluster can be chosen manually in RecoverPoint.
Alternatively, the preferred RecoverPoint cluster can be set to follow VPLEX bias rules
for the VPLEX consistency group.
l If the preferred RecoverPoint cluster is not at the VPLEX winning site, a production
switchover occurs. During the site departure, RecoverPoint at the VPLEX winning site
continues to replicate to the remote site.
l The losing site does not receive any I/Os during the site departure and does not
maintain an up-to-date journal.
l A local copy at the losing site will not be updated.
executed, active-active access to the volume on both Metro sites becomes available
immediately, as the remote copy is now the production copy.
Figure 70 MetroPoint four-site configuration
Data is not replicated to all four sites. At a consistency group level, it is replicated to a
single target. If the target replica is a distributed RAID 1, it will be in a fractured state.
If production fails over from across regions, the target fractured volume will automatically
start rebuilding the mirror leg. This provides continuous availability between the two
sites, once the distributed volume is rebuilt and synchronized.
The four-site MetroPoint topology provides regional replication protection for two regions
or two VPLEX Metro systems.
In the four-site configuration, production applications running in Site A and Site B are
replicated to Site C for disaster protection. Distributed volumes at VPLEX Metro Site C and
Site D can be replicated to a single target volume at remote Site A or Site B (but not both).
Different applications, with volumes in different consistency groups, can run in Site C and
Site D as active and standby source sites and be protected in one remote site, Site A or
Site B.
Note
Data is not replicated to all four sites. Each MetroPoint group is replicated to a single
target. Different MetroPoint groups can have different targets.
If a regional disaster causes both sides of the Metro to fail, operations can fail over to the
disaster recovery Site C.
When the MetroPoint group fails over to Site C, the replication will be from Site C to Metro
Site A or Site B, whichever is currently defined as the winner site in VPLEX.
The production source is now in the remote site and one of the Metro sites becomes the
replica. The mirrored leg of the fractured DR1 is rebuilt at the remote site.
Figure 73 During failover, Site C and Site D become the production sites
This figure illustrates that production sites (A and B) have failed over to the remote Site C,
which is now the production site with Site D. The fractured distributed RAID 1 volume on
Site C changes from a replica volume to a production volume, so Site D is rebuilt. The DR1
volume on Site A and Site B becomes fractured during the failover, as it is now the replica
copy at the remote site for production sites C and D.
In a recovery following Site A and Site B failure, Site C will restore the last active site, Site
A or Site B, which will automatically become the winner site and be restored. VPLEX
rebuilds the non-winner site. Site A and Site B are restored as the production sites and
the DR1 volume at Site C becomes fractured again, as it is now the replica copy.
Upgrading to MetroPoint
Install a new MetroPoint configuration, or upgrade from existing VPLEX and/or
RecoverPoint configurations easily and non-disruptively.
Note
All operating system platforms and clusters normally supported with VPLEX are fully
supported with MetroPoint. See the latest copy of Simple Support Matrix for VPLEX for
specific details.
The following MetroPoint deployments are available:
l New MetroPoint installation- Install VPLEX Metro and RecoverPoint and create
MetroPoint groups.
l Upgrade from RecoverPoint- Non-disruptively upgrade from RecoverPoint to a
MetroPoint topology by installing VPLEX into an existing RecoverPoint environment.
Attach RecoverPoint to both VPLEX clusters.
l Upgrade from VPLEX- Non-disruptively upgrade to a MetroPoint topology from an
existing VPLEX Local or Metro configuration. This adds a third site to provide
RecoverPoint local protection on both local sites and provides remote protection at
the distant site. Attach RecoverPoint at the second VPLEX Metro site.
l Upgrade from an existing VPLEX and RecoverPoint configuration- Non-disruptively
upgrade to MetroPoint from an existing VPLEX and RecoverPoint configuration that is
currently using remote replication to a third site. Attach RecoverPoint at the second
VPLEX Metro site and convert existing RecoverPoint consistency groups to MetroPoint
groups.
See the VPLEX procedures in the EMC SolVe Desktop for more information about
installing and upgrading to MetroPoint.
MetroPoint groups
After MetroPoint has been enabled on the system, create new MetroPoint groups or
convert existing RecoverPoint consistency groups to MetroPoint groups to enable
MetroPoint protection. All MetroPoint groups are protected at both sides of the VPLEX
Metro by RecoverPoint.
A MetroPoint group can contain the following copies:
l One production copy at the active source site
l One production copy at the standby source site
l Zero, one, or two local copies
l One remote copy
A MetroPoint group, when configured with two local copies, will contain the maximum of
five copies.
MetroPoint groups and regular RecoverPoint consistency groups can co-exist in the same
system. However, regular RecoverPoint consistency groups are protected on one side
only, at the VPLEX preferred site.
A MetroPoint group consists of distributed volumes or local volumes. They cannot be
combined in the same MetroPoint group. One MetroPoint group can replicate to a remote
site and another MetroPoint group can replicate to a different remote site.
Each MetroPoint group can be configured to:
l Replicate from Metro Site A
l Replicate from Metro Site B
l Follow VPLEX bias rules
The best practice is to select the default "Follow VPLEX bias rules" which provides added
protection in RecoverPoint that will select the side that the VPLEX consistency group will
declare as the winner in the instance of link failures in the VPLEX Metro system and
prevent any unnecessary swapping of the source of replication.
See the EMC RecoverPoint Administration Guide and VPLEX procedures in the SolVe
Desktop for more information on creating or converting a consistency group to a
MetroPoint group.
Creating a MetroPoint group
A MetroPoint group protects distributed RAID 1 volumes on both sides of the VPLEX
Metro.
To create a new MetroPoint group:
l In Unisphere for VPLEX, ensure you have provisioned storage and created the
distributed devices you want to use.
l Create VPLEX consistency groups that are RecoverPoint enabled.
l Expose the storage in the VPLEX consistency groups to RecoverPoint and expose the
hosts through the storage views.
l In Unisphere for RecoverPoint, ensure that the distributed device is exposed to the
RecoverPoint cluster local to each VPLEX Metro cluster and it belongs to a storage
view on VPLEX Metro Site A and Metro Site B. RecoverPoint will see the same volume
on both Metro sites and use it to filter distributed volumes.
Converting to a MetroPoint group
An existing consistency group can be converted to a MetroPoint group.
There are two ways to convert existing consistency groups:
l Take a RecoverPoint consistency group that is already using distributed RAID 1
volumes for production volumes and convert it to a MetroPoint group.
l Upgrade from a consistency group that is using local volumes by converting the local
volumes to VPLEX distributed volumes and creating a MetroPoint group.
Currently, upgrading local volumes is disruptive to RecoverPoint replication. The volumes
that are to be converted from local devices to distributed devices must be removed from
RecoverPoint replication sets and VPLEX consistency groups. Host I/O continues.
Currently VPLEX architecture constraints prevent adding mirror legs to volumes that are
splitting I/O to RecoverPoint, and VPLEX also does not allow different volume topologies
to exist in the same consistency group.
The local volumes are removed from the RecoverPoint replication set, then from the
VPLEX consistency group. A mirror leg is added to create a VPLEX distributed device,
which is then placed back in a RecoverPoint enabled consistency group. Create
replication sets for the volumes in the RecoverPoint consistency group in the Unisphere
for RecoverPoint user interface, then create the MetroPoint group using the Protect
Volumes Wizard in the same user interface.
Failover
The following are the two types of failover:
l Production failover - This failover refers to the act of reversing replication. In a
production failover, the original production source becomes the replica target and the
earlier replica target becomes the replication source.
l Active source failover - This failover is unique to MetroPoint groups having two
production sources, one active and one standby for a given remote replication target.
During active source failover, the active source and standby source will exchange
their roles, that is, the active source becomes the standby source and the standby
source becomes the active source.
The VPLEX SolVe Desktop provides more information on performing a failover.
The following figure illustrates source failover. If the active source site (Site A) fails, the
best-site mechanism initiates a source failover. The delta marker information from the
standby site (Site B) is resynchronized to the remote site (Site C). Site B becomes the
active source site.
You can also manually failover the active source site. Manually failing over the active
source of production depends upon the group policy of the RecoverPoint consistency
group and the set preferred cluster. The preferred cluster can be set to either a specific
RecoverPoint cluster, or to follow VPLEX bias rules.
l If the preferred cluster is set to Follow VPLEX bias rules, then the active
source site follows the VPLEX winner rules set for the consistency group. If you want
to swap from Site A to Site B, you must change the VPLEX winner rules to specify that
Site B wins instead of Site A and the RecoverPoint active source automatically swaps
to the other VPLEX site.
l If the preferred cluster in Recoverpoint is set to a specific RecoverPoint cluster, the
active source of replication can be swapped to the other site by selecting the
opposite RecoverPoint cluster.
The VPLEX SolVe Desktop provides more information on performing a manual switch over
of the active site.
Note
If a distributed RAID 1 volume is selected as a local copy, then that distributed RAID 1
volume must have its configured bias set to the site where it will be a local copy. I/O is
replicated to only one leg of the distributed RAID 1 and is not distributed to the other leg
of the distributed RAID 1 copy. However, after production failover to the distributed RAID
1 local copy, the host can access the distributed RAID 1 from both VPLEX sites and VPLEX
rebuilds the distributed RAID 1 legs to synchronize both legs of the distributed RAID 1
volume.
Recover production
The MetroPoint topology enables you to perform recover production from each of the
copies (either remote or local on any site) without losing the journal on either of the
Metro sites.
The following figure illustrates the recover production flow. In this example, the recover
production happens from the remote copy. The remote copy is moved to the image
access state. The active source site (Site A) is moved to FAIL-ALL, which stops host I/O
and VPLEX I/O synchronous distribution between Site A and Site B (distributed RAID 1
fracture). In this state, only replication I/O from RPA is allowed on the Distrubuted RAID 1
leg at Site A. The replication to the local copy is paused. The splitters in Site B are in the
Marking On Hosts (MOH) state because of a mirror fracture. An init is performed from the
remote copy to Site A. Then the replication direction is reversed back and Site A is out of
FAIL-ALL, mirror recovery of Site B starts with VPLEX distributed RAID 1 rebuild using Site
A as the source. When mirror recovery ends, the local copy is initialized. Finally, the sync-
peer-splitter removes the redundant information and information from Site B to Site C is
synchronized.
Figure 77 Recover production
ON ON
I I
O O
OFF OFF
ON ON
I I
O O
OFF OFF
ON ON
I I
O O
OFF OFF
OFF OFF
O O
I I
ON ON
OFF OFF
Director 1B
O O
I I
ON ON
Engine 1
Director 1A
SPS 1
VPLX-000215
ON ON
I I
O O
OFF OFF
ON ON
I I
O O
OFF OFF
ON ON
I I
O O
OFF OFF
Director 2B
Engine 2
OFF
O
I
ON
OFF
O
I
ON
Director 2A
SPS 2
OFF OFF
Director 1B
O O
I I
ON ON
Engine 1
Director 1A
SPS 1
VPLX-000214
Director 4B
Engine 4 Director 4A
ON ON
I I
O O
OFF OFF
SPS 4
ON
I
ON
I
Director 3B
Engine 3
O O
OFF OFF
Director 3A
SPS 3 ON
I
O
OFF
ON
I
O
OFF
Director 2B
Engine 2
OFF
O
I
ON
OFF
O
I
ON
Director 2A
SPS 2
OFF OFF
Director 1B
O O
I I
ON ON
Engine 1
Director 1A
SPS 1
VPLX-000213
VS1 engine
Figure 81 VS1 engine
Note
The WAN COM ports on IOMs A4 and B4 are used if the inter cluster connections are over
Fibre Channel, and the WAN COM ports on IOMs A5 and B5 are used if the inter cluster
connections are over IP.
In the above figure, IOMs A2 and B2 each contain four Fibre Channel ports for VPLEX
Metro WAN connection.
VPLEX Metro also supports IP WAN connections, in which case these IOMs each contain
two Ethernet ports.
In a VPLEX Local configuration, these IOMs contain no ports.
Engine 4: Engine 4:
Director 4B 128.221.253.42 Director 4B 128.221.252.42
Director 4A 128.221.253.41 Director 4A 128.221.252.41
Engine 3: Engine 3:
Director 3B 128.221.253.40 Director 3B 128.221.252.40
Director 3A 128.221.253.39 Director 3A 128.221.252.39
FC switch B 128.221.253.34
Service port
Public Ethernet port
128.221.252.2
Customer-assigned
FC switch A 128.221.252.34
Engine 2: Engine 2:
Director 2B 128.221.253.38 Director 2B 128.221.252.38
Director 2A 128.221.253.37 Director 2A 128.221.252.37
Engine 1: Engine 1:
Director 1B 128.221.253.36 Director 1B 128.221.252.36
Director 1A 128.221.253.35 Director 1A 128.221.252.35
VPLX-000107
Engine 4: Engine 4:
Director 4B 128.221.253.74 Director 4B 128.221.252.74
Director 4A 128.221.253.73 Director 4A 128.221.252.73
Engine 3: Engine 3:
Director 3B 128.221.253.72 Director 3B 128.221.252.72
Director 3A 128.221.253.71 Director 3A 128.221.252.71
FC switch B 128.221.253.66
Service port
Public Ethernet port
128.221.252.2
Customer-assigned
FC switch A 128.221.252.66
Engine 2: Engine 2:
Director 2B 128.221.253.70 Director 2B 128.221.252.70
Director 2A 128.221.253.69 Director 2A 128.221.252.69
Engine 1: Engine 1:
Director 1B 128.221.253.68 Director 1B 128.221.252.68
Director 1A 128.221.253.67 Director 1A 128.221.252.67
VPLX-000108
All cluster sizes, VPLEX Metro only Fibre Channel WAN Figure 96 on page 139
COM
Management server
Purple, 71 in.
Green, 37 in.
Engine 1
VPLX-000052
Engine 1
12 in. 12 in.
VPLX-000069
Engine 1
39 in.
(2 cables)
VPLX-000063
Note
Both Fibre Channel cables are light blue. However, the A side cable has a blue label,
and the B side cable has an orange label.
OFF
O
OFF
O
Management server
I I
ON ON
OFF OFF
O O
I I
ON ON
OFF OFF
O O
I I
ON ON
Engine 1
SPS
VPLX-000041
Green, 48 in.
Fibre Channel switch A
Purple, 71 in.
Management server
Purple, 71 in.
Green, 37 in.
Engine 2
Green, 20 in.
Purple, 20 in.
Engine 1
VPLX-000043
UPS B
UPS A
40 in.
40 in.
Engine 2
12 in. 12 in.
Engine1
12 in. 12 in.
VPLX-000067
Engine 2
Engine 1
VPLX-000057
Note
All 16 Fibre Channel cables are light blue. However, the cables connected to Fibre
Channel switch A have blue labels, and the cables connected to switch B have orange
labels.
UPS B
UPS A
Management server
I I
Engine 2
O
I O
I
SPS 2
O
I O
I
Engine 1
SPS 1
VPLX-000042
Engine 4
Green, 20 in.
Purple, 20 in.
Green, 48 in.
Engine 3
Purple, 71 in.
Management server
Green, 37 in.
Engine 2
Purple, 71 in.
Green, 20 in.
Purple, 20 in.
Engine 1
VPLX-000044
Engine 4
12 in. 12 in.
Engine 3
12 in. 12 in.
UPS B
UPS A
40 in. 40 in.
Engine 2
12 in. 12 in.
Engine1
12 in. 12 in.
VPLX-000065
Engine 4
Engine 3
79 in.
(all 16 cables) Fibre Channel switch B
Engine 2
Engine 1
VPLX-000055
Note
All 16 Fibre Channel cables are light blue. However, the cables connected to Fibre
Channel switch A have blue labels, and the cables connected to switch B have orange
labels.
ON
I
ON
I
Engine 4
O O
OFF OFF
SPS 4
ON ON
I I
O O
OFF OFF
Engine 3
ON
I
O
OFF
ON
I
O
OFF
SPS 3
UPS B
UPS A
OFF
O
OFF
O
Management server
I I
ON ON
OFF OFF
Engine 2
O O
I I
ON ON
SPS 2
OFF OFF
O O
I I
ON ON
Engine 1
SPS 1
VPLX-000211
Intercluster Intercluster
ISL 2
COM SAN COM SAN
switch 1B switch 2B
VPLX-000317
IP
subnet B
VPLX-000368
AccessAnywhere The technology that enables VPLEX/VE sites to provide access to information between
sites that are separated by distance.
alarm (vSphere) - A vSphere alarm is an entity that monitors one or more properties of a virtual machine,
such as CPU load. Alarms send notifications as directed by the configurable alarm
definition.
array A collection of disk drives where user data and parity data may be stored.
back-end IP SAN (Array- The IP network that connects physical storage from the iSCSI storage arrays to the
facing) vDirectors. The vDirectors consume the physical storage from the array and produce
distributed virtual volumes.
backend port The virtual adapter port used by the vDirector to connect to the back-end (Array-facing) IP
SAN. This port acts as an iSCSI initiator.
cache Temporary storage for recent writes and recently accessed data. Disk data is read through
the cache so that subsequent read references are found in the cache.
cache coherency Managing the cache so data is not lost, corrupted, or overwritten. With multiple
processors, data blocks may have several copies, one in the main memory and one in each
of the cache memories. Cache coherency propagates the blocks of multiple users
throughout the system in a timely fashion, ensuring the data blocks.
disaster recovery (DR) The ability to restart system operations after an error, preventing data loss.
distributed datastore A virtual storage volume that is backed by physical storage from two sites within a vSphere
stretched cluster. The distributed virtual volumes is the backing device for a distributed
datastores.
distributed virtual The globally visible storage volumes in VPLEX/VE. The distributed virtual volumes provide
volumes the storage for creating distributed datastores.
front-end IP SAN (ESXi- The IP network that presents distributed virtual storage from the vDirectors to ESXi hosts.
facing) Distributed virtual storage is used to provision vSphere distributed datastores.
front end port - The virtual adapter port used by the vDirector to connect to the front-end (ESXi-facing) IP
SAN. This port acts as an iSCSI target.
input/output (I/O) Any operation, program, or device that transfers data to or from a computer.
Inter-Site Management The virtual private network (VPN) tunnel between the management servers across sites. For
Network high availability, there are two such networks between sites.
iSCSI The internet small computer system interface (iSCSI) protocol that allows commands to
travel through IP networks, which carries data from storage units to servers anywhere in a
computer network.
LAN A local area network (LAN) is a group of computers and associated devices that share a
common communications line and typically share the resources of a single processor or
server within a small geographic area.
load balancing Distributing the processing and communications activity evenly across a system or
network so no single device is overwhelmed. Load balancing is especially important when
the number of I/O requests issued is unpredictable.
local COM - Communication between vDirectors within a site; note that Local COM and Management
COM share the Local COM (Intra-Site) Network.
Local COM Network (Intra- The private IP networks connecting all VPLEX/VE virtual machines within one site. For high
Site Network) availability, there are two such networks at each site. This is a private network. The virtual
switches for this network are selected during installation, but the IP addresses are static
and cannot be controlled by the installer.
logging volume Logging volumes keep track of blocks written during an inter-site link outage or when one
leg of a DR1 becomes unreachable. After the inter-site link or leg is restored, the VPLEX/VE
system uses the information in the logging volumes to synchronize the mirrors by sending
only the changed blocks across the link.
logical unit number (LUN) Virtual storage to which a given server with a physical connection to the underlying storage
device may be granted or denied access. LUNs are used to identify SCSI devices, such as
external hard drives, connected to a computer. Each device is assigned a LUN number
which serves as the unique address of the device.
management COM Communication between the vManagementServer and the vDirectors. Management COM
and Local COM share transport on the Local COM (Intra-site) Network.
meta volume A storage volume used by the system that contains the metadata for all the virtual volumes
managed by the system. There is one metadata storage volume per site.
meta-volume backup A meta-volume backup contains the backup copies of the metadata files of a VPLEX/VE
system. You can specify a schedule for running a meta-volume backup. There is one
metadata volume per site.
migration (VPLEX/VE) Migrate one of the backing devices for a distributed datastore to a new storage device with
no application downtime. See the action menu for a distributed datastore in the vCenter
Inventory. Also called device migration.
migration (vSphere) For vSphere, migration is the process of moving a virtual machine between hosts. Unless
vMotion is used, the virtual machine must be powered off when you migrate it. See also
migration with vMotion.
migration with vMotion For vSphere, this is the process of moving a virtual machine that is powered on and meets
(vSphere) selected requirements, including the activation of vMotion on both the source and target
hosts. When you migrate a virtual machine using vMotion, the operations of the virtual
machine can continue without interruption.
open virtual appliance A packaging format for virtual machines that allows virtual machine templates to be
(OVA) distributed, customized, and instantiated on any OVA supporting host. VPLEX/VE is
distributed as an OVA file.
Open Virtualization - A distribution format for virtual appliances that uses existing packaging tools to combine
Format (OVF) one or more virtual machines with a standards-based XML wrapper. OVF gives the
virtualization platform a portable package containing all required installation and
configuration parameters for virtual machines. This format allows any virtualization
platform that implements the standard to correctly install and run virtual machines.
VPLEX/VE uses the OVF standard packaging format.
physical network A network of physical machines (plus cabling, switches, routers, and so on) that are
connected so that they can send data to and receive data from each other. See also virtual
network.
port group (vSphere) A construct for configuring virtual network options such as bandwidth limitations and VLAN
tagging policies for each member port. Virtual networks that are connected to the same
port group share network policy configuration. See also virtual network, VLAN (virtual local
area network).
provisioning Virtual The process of creating a functioning virtual machine by assigning resources such as CPU,
Machines memory, and virtual hardware and then deploying a system image.
RAID The use of two or more storage volumes to provide better performance, error recovery, and
fault tolerance.
rebuild The VPLEX/VE process of reconstructing data onto a spare or replacement drive after a
drive failure. Data is reconstructed from the data on the surviving disks, assuming
mirroring has been employed.
Site-bias Site at which I/O will continue during an inter-site failure. Each datastore must have
separate site bias settings.
subnet A logical grouping of connected network devices. The devices that are part of a subnet
normally belong to a local area network.
vDS (vNetwork A vSphere virtual switch with an abstract representation of multiple hosts sharing the
Distributed Switch) same vSwitch (same name, same network policy) and port group. The vDS configuration
resides on the vCenter Server, instead of residing on an individual ESXi host. This
abstraction allows virtual machines to vMotion among multiple hosts with no concern for
the switch configuration.
virtual appliance - A software solution that is composed of one or more virtual machines. A virtual appliance
is packaged as a unit by an appliance vendor and is deployed, managed, and maintained
as a unit.
virtual application (vApp) A VMware specific term for a virtual appliance delivered in an OVF package and deployed
to a vCenter Server. The VMware vApp is a self-describing software package composed of
one or more virtual machine.
virtual director (vDirector) The core component in a VPLEX/VE system. A VPLEX/VE vDirector is responsible for
providing the virtual-to-physical I/O translations for vSphere distributed datastores. There
number vDirectors in each VPLEX/VE site is determined by the license.
virtual local area network A logical group of workstations, servers and network devices that appear to be on the
(VLAN) same local area network (LAN) even though they are geographically distributed. A VLAN
allows a network of computers and users to communicate in a simulated environment as if
they exist in a single LAN and are sharing a single broadcast and multicast domain.
virtual machine A virtual machine is a software computer that, like a physical computer, runs an operating
system and applications. Multiple virtual machines can operate on the same host system
concurrently.
virtual machine A file containing a virtual machine configuration; a specification of which virtual devices,
configuration file (vmx such as disks and memory, are present in a virtual machine and how they are mapped to
file) host files and devices. This .vmx file is created when you create the virtual machine. It is
used to identify and run a specific virtual machine.
virtual management The management server virtual machine that contains the interface to manage the virtual
server directors and site-level controls for a VPLEX/VE system.
(vManagementServer)
virtual network A network connecting virtual machines that does not depend on physical hardware
connections. For example, you can create a virtual network between a virtual machine and
a host that has no external network connections.
virtual switch A virtualized network switch used by ESXi to manage traffic between virtual machines and
the physical network adapters on the ESXi host.
VLAN (virtual local area A software-managed logical segmentation of a physical LAN. Network traffic within each
network) segment is isolated from traffic in all other segments.
VMFS (Virtual Machine A file system that is optimized for storing virtual machines. One VMFS partition is
File System) supported for each SCSI storage device or LUN.
VMkernel In vSphere ESXi, a high-performance operating system that occupies the virtualization
layer and manages most of the physical resources on the hardware, including memory,
physical processors, storage, and networking controllers.
VMware ESXi A virtualization layer run on physical servers that abstracts processor, memory, storage,
and resources into multiple virtual machines. A VMware vSphere feature.
VMware ESXi cluster An ESXi cluster is a group of ESXi hosts. A cluster is an entity that manages the resources
of all the hosts in it. To deploy VPLEX/VE, you must have an ESXi cluster with vSphere High
Availability (HA) and vSphere Distributed Resource Scheduler (DRS) features enabled.
VMware ESXi host An ESXi host is a computer that uses the VMware ESXi virtualization software. The hosts
provide the physical resources such as the CPU and memory that the virtual machines use
to connect to the storage and the network.
VMware ESXi host An ESXi host is a computer that uses the VMware ESXi virtualization software. The hosts
provide the physical resources such as the CPU and memory that the virtual machines use
to connect to the storage and the network.
VMware vCenter Server The central point for configuring, provisioning, and managing virtualized IT environments.
It provides essential datacenter services such as access control, performance monitoring,
and alarm management. VPLEX/VE runs on vCenter Server version 5.1 or later.
VMware vCenter Server The vCenter Server is the centralized management tool for the VMware vSphere suite. As a
single management application, the vCenter Server manages and monitors the multiple
ESXi servers and virtual machines that are part of different ESXi servers. To deploy and
configure VPLEX/VE, you must use the vCenter Server version 5.1 or later.
VMware vMotion VMware vMotion enables the live migration of running virtual machines from one physical
server to another with zero downtime, continuous VMware availability, and complete
transaction integrity. VMotion is a key enabling technology for creating the dynamic,
automated, and self-optimizing datacenter.
VMware vMotion - VMware VMotion enables the live migration of running virtual machines from one physical
server to another with zero downtime, continuous VMware availability, and complete
transaction integrity. VMotion is a key enabling technology for creating the dynamic,
automated, and self-optimizing datacenter.
VMware vSphere client The vSphere client communicates to vCenter Server and enables you to perform the actions
that you want to do on your vSphere environment. You can either download this client from
the VMware website or configure the vSphere web client on the vCenter Server.
VMware vSphere Client An interface that enables users to connect remotely to vCenter Server or ESXi from any
(for Windows) Windows PC. Also known as vSphere Desktop Client or '.Net' client. The VPLEX/VE
Management Extension (plug-in) modules are available only in the vSphere Web Client, not
in the vSphere Windows Client.
VMware vSphere Platform Platform for virtualization which manages large collections of infrastructure objects, such
as compute, storage and networking, as a seamless and dynamic operating environment.
The VMware vSphere platform includes:
l VMware ESXi
l VMware vCenter Server
l VMware vSphere Web Client
l vSphere Virtual Machine File System (VMFS)
l vSphere vMotion
l vSphere Storage vMotion
l vSphere High Availability (HA)
l vSphere Distribute Resource Scheduler (DRS)
l vSphere Distributed Switch (VDS)
VMware vSphere Web A Web interface that enables users to connect remotely to vCenter Server from a variety of
Client Web browsers and operating systems.
vNIC A virtual network interface card that is configured on top of a system's physical Network
adapter. See also NIC (network interface card).
VPLEX/VE Manager The name for the vSphere management extension for VPLEX/VE.
VPLEX/VE Manager for The name for the vSphere management extension for VPLEX/VE.
vSphere
vSphere cluster A vSphere cluster is a group of ESXi hosts. A cluster is an entity that manages the
resources of all the hosts in it. To deploy VPLEX/VE, you must have an vSphere cluster with
vSphere High Availability (HA) and vSphere Distributed Resource Scheduler (DRS) features
enabled.
vSphere Distributed A virtual switch that can span multiple ESXi hosts, enabling significant reduction of on-
Switch (VDS) going network maintenance activities and increasing network capacity. This increased
efficiency enables virtual machines to maintain consistent network configuration as they
migrate across multiple hosts.
vSphere Distribute Allocates and balances computing capacity dynamically across collections of hardware
Resource Scheduler (DRS) resources for virtual machines.
vSphere High Availability A feature that provides high availability for virtual machines. If a server fails, affected
(HA) virtual machines are restarted on other available servers that have spare capacity.
vSphere Storage vMotion Enables the migration of virtual machine files from one datastore to another without
service interruption. You can place the virtual machine and all its disks in a single location,
or select separate locations for the virtual machine configuration file and each virtual disk.
The virtual machine remains on the same host during Storage vMotion. Migration with
Storage vMotion lets you move the virtual disks or configuration file of a virtual machine to
a new datastore while the virtual machine is running. Migration with Storage vMotion
enables you to move a virtual machine's storage without any interruption in the availability
of the virtual machine.
vSphere stretched cluster A vSphere stretched cluster contains compute, network and storage resources at two
different physical locations called sites. The sites may be in the same building or different
physical locations with synchronous distances. The sites act as fault domains; a
component failure in one site cannot affect the other site. The stretched cluster ensures
high availability and disaster recovery capabilities of the sites.
VPLEX/VE must be deployed on a vSphere stretched cluster.
Also called ESXi stretched cluster or Metro Stretched Cluster (vMSC).
vSphere Virtual Machine A high performance cluster file system for ESXi virtual machines.
File System (VMFS)
vSphere vMotion Enables the migration of powered-on virtual machines from one physical server to another
with zero down time, continuous service availability, and complete transaction integrity.
Migration with vMotion cannot be used to move virtual machines from one datacenter to
another.
vSS (vNetwork Standard - A virtual switch that is configured on an individual ESXi host. To ensure operations like
Switch or vSwitch) vMotion across ESXi hosts, an administrator must manually maintain consistency of the
vSwitch configuration across the ESXi hosts.
WAN COM (Inter-Site The IP network connecting the vDirectors across sites. As a best practice, this should be a
Network) private network. For high availability, there are two such networks at each site.
zoning Provides access control in the SAN topology. Zoning defines which HBAs can connect to
which storage processors. You can have multiple ports to the same storage processor in
different zones to reduce the number of presented paths.
I/O failure 41 O
I/O failures 40 outages 64
I/O module failures 40
I/O paths 37
integrated storage 51 P
integrity 64 passwords 79
intra-cluster IP 41 path optimization 74
IPsec 79 performance 27, 56, 59, 77, 78
port failures 39
port usage 79
J power 37, 38
journal volume 95 power failure 76
preface 11
L production copy 114
latency 86 production failover 116
LDAP 51, 79 production site 94
limits 59 production volume 95
link failure 42 provisioning 52, 53
load balancing 59, 77 provisioning storage 51
local and remote protection 98 provisioning wizard 52
local copies 114
local copy 118 Q
local mirrors 57 quad-engine cluster 31, 35
local protection 9799 quorum 65
local replication 96, 97, 99
local visibility 60
log splitting 27 R
logged access 95 recover production 118
logging volumes 38, 58, 75, 76 Recover production 118
LUN mapping 39 RecoverPoint 94, 9799, 102
LUN masking 39 RecoverPoint appliance 95
RecoverPoint Appliance 94, 99, 101, 118
RecoverPoint cluster 100, 101
M RecoverPoint configurations 96
management from clusters 49 RecoverPoint consistency group 94, 95, 116
Management GUI 49 RecoverPoint failover 95
management server 30, 43 RecoverPoint splitter 94, 100
management server failure 43 RecoverPoint Volumes 95
metadata volumes 75 recovery 104
Metro HA 87, 88, 90 redundancy 3743, 56, 57, 64, 6668, 94
MetroPoint 103, 114, 116118 redundant connections 67
MetroPoint consistency group 117 redundant directors 66
MetroPoint group 117 redundant fabrics 67
MetroPoint groups 114 redundant paths 66, 68
MetroPoint upgrade 114 related documentation 11
migration 82, 84 reliability 39, 42, 43, 53, 58
mirror recovery 118 remote copies 114
mirroring 57 remote copy 117
mirrors 21, 38, 39, 58, 75 remote mirrors 57
mobility 19, 8285 remote protection 97
monitoring 38, 50, 51, 7779 remote replication 96, 99
monitoring CLI 78 replica site 94
monitors 78 replica volume 95
MSCE 98 replica volumes
multi-pathing 64, 67, 68 rollback 97
multipathing replication 94, 99, 101104, 110
SFP 39 repository volume 95
resiliancy 59
N resilience 21, 24, 41, 75
network failures 72 resiliency 72
non-disruptive upgrade 26 resilliance 68
REST 50
rollback 95 U
RPA 99 uninterupted power supply 38
RPO 69 Unisphere for VPLEX 49
RTO 69, 72, 87, 90 Unisphere GUI 49
RTT 87 Unisphere monitoring tool 77
rule sets 69 upgrade 26, 114
UPS 38
S user roles 79
scalability 21
security 51, 79 V
single-engine cluster 31, 33 vault 76
site distribution 68 vCenter 102
site recovery 102 VIAS 51
Site Recovery Manager 102 virtual access 95
SMI-S 51 virtual volumes 57, 67
SNMP 50 visibility 55, 59, 60, 86
splitter 100 VMware cross-cluster connection 74
SPS 38 VMware HA 87
SRM 102 VPLEX Metro failures 72
standby power 38 VPLEX Witness 24, 42, 6872, 74, 87, 88, 90
standby production copy 114
statistics 78
storage 51 W
storage arrays 56 WAN connectivity 36
subnets 41 WAN link load 77
support information 11 WWN 22
synchronous consistency groups 59
Z
T zoning 37
technology refresh 82
thick rebuilds 53
thin rebuilds 53