Documente Academic
Documente Profesional
Documente Cultură
Abstract
A solution consisting of ProtectPoint and VPLEX that
allows file system backup and restore to take place
between VMAX3 and Data Domain within the virtual
storage infrastructure. This capability not only reduces
host I/O and CPU overhead, allowing the host to focus on
processing and applications, but also provides higher
efficiency for the backup and recovery process itself
whilst allowing VPLEX to non-disruptive mobility and
continuous availability.
Copyright 2015 EMC Corporation. All Rights Reserved.
EMC believes the information in this publication is accurate of its publication date. The
information is subject to change without notice.
The information in this publication is provided as is. EMC Corporation makes no
representations or warranties of any kind with respect to the information in this
publication, and specifically disclaims implied warranties of merchantability or fitness for
a particular purpose.
Use, copying, and distribution of any EMC software described in this publication requires
an applicable software license.
For the most up-to-date listing of EMC product names, see EMC Corporation Trademarks on
EMC.com.
All other trademarks used herein are the property of their respective owners.
Part Number H14720
EMC ProtectPoint addresses these gaps by enabling best in class EMC products -- VMAX3 storage array and Data
Domain system to provide storage-based backup and recovery in an automated, efficient, and integrated
manner.
The enabling technology for the ProtectPoint is the integration between VMAX3 and Data Domain systems which
allows file system backup and restore to take place entirely within the storage infrastructure! This capability not
only reduces host I/O and CPU overhead, allowing the host to focus on servicing applications and transactions,
but also provides higher efficiency for the backup and recovery process itself.
Backup efficiencies are introduced by not requiring any read or write I/Os of the data files by the host. Instead,
ProtectPoint creates an internal point-in-time consistent copy of the database, and then copies it directly to the
Data Domain system. ProtectPoint for File Systems allows for scripting to carry out any host side activities to
ensure an application consistent backup is taken. As soon as the snapshot is created, the file system is
operation returned to normal. The file system snapshot is then incrementally copied to the Data Domain system
in the background, while file system operations continue as normal. The combination of ProtectPoint, VMAX3,
and Data Domain allows the highest backup efficiency.
Restore efficiencies are introduced in a similar way by not requiring any read or write I/Os of the data files by the
host. Instead, Data Domain places the required backup-set on its restore encapsulated devices. The restore
devices can be directly mounted to a Mount host for small-scale data retrievals, or mounted to the Production
host and cataloged with RMAN so the full scale of RMAN functionality can be used for production-database
recovery (e.g. fixing physical block corruption, missing data files, etc.). A 3rd option is available, where the
restore devices content is copied by SnapVX, overwriting the native VMAX3 Production devices. This option is
best used when the production file system requires a complete restore from backup, or for a large-scale
recovery that shouldnt be performed from the encapsulated Data Domain devices.
Audience
This white paper is intended for technology architects, storage administrators, and system administrators who
are responsible for implementing, managing, and maintaining file system backup and recovery strategy with
VMAX3 storage systems. It is assumed that readers have some familiarity with VPLEX and the EMC VMAX3
family of storage arrays, and are interested in achieving improved file system availability, performance, and
ease of storage management.
Introduction
This white paper addresses following administrative tasks for ProtectPoint operations consisting of VPLEX,
ProtectPoint for File Systems, VMAX3, and DataDomain:
Pre-backup checks
ProtectPoint for File Systems backups
Pre-recovery checks
ProtectPoint for File Systems file recovery
This paper examines the technical impact of VPLEX on ProtectPoint for VMAX3 using ProtectPoint for File
Systems backup and recovery. The necessary pre-backup considerations, logical checks, process changes and
use case examples are provided. The examples illustrate the correct way to account for VPLEX and potential
storage conditions that could impact the consistency of backup.
Term Description
Application Consistent vs A snapshot or backup is only considered Application Consistent if that snapshot or backup is
Application Aware taken in a way specifically supported by a given application. This may require taking a given
application in and out of backup mode. A snapshot or backup is considered to be application
aware if the snapshot of backup is taken through direct integration with a given applications API.
Storage Consistent vs. File systems make a distinction between a crash consistent and fully consistent state of the file
Crash Consistent file system. A fully consistent state requires all file system journal data, file data, and file metadata
system to be consistent (see storage consistent replications) on the storage media. In this state, file
systems can be simply mounted without user intervention. Crash consistent state, on the other
hand, requires file system recovery; often times relying on the file system journal and/or logging
to achieve file system consistency before the file system can be mounted by the host.
RTO and RPO Recovery Time Objective (RTO) refers to the time it takes to recover a database after a failure.
Recovery Point Objective (RPO) refers to any amount of data-loss after the recovery completes,
where RPO=0 means no data loss of committed transactions.
Storage consistent Storage consistent replications refer to storage replication (local or remote) that maintains write-
replications order fidelity at the target devices, even while the application is running. To the host, the
replicated file system image is crash consistent.
VMAX Federated Tiered Federated Tiered Storage (FTS) is a feature of VMAX3 that allows an external storage to be
Storage (FTS) connected to the VMAX3 backend and provide physical capacity that is managed by VMAX3
software.
VMAX3 HYPERMAX OS HYPERMAX OS is the industrys first open converged storage hypervisor and operating system. It
enables VMAX3 to embed storage infrastructure services like cloud access, data mobility and
data protection directly on the array. This delivers new levels of data center efficiency and
consolidation by reducing footprint and energy requirements. In addition, HYPERMAX OS delivers
the ability to perform real-time and non-disruptive data services.
Storage Groups Storage Groups can be used to (a) present devices to host (LUN masking) on VMAX3, (b) specify
FAST Service Levels (SLOs) to a group of devices on VMAX3, and (c) manage groups of devices
used by VPLEX and/or for replication software such as VMAX3 SnapVX and SRDF.
VMAX3 Storage Groups can be cascaded, such as the child storage groups can be used for
setting FAST Service Level Objectives (SLOs) and the parent used for LUN masking of all the
database devices to the host.
VMAX3 TimeFinder Previous generations of TimeFinder referred to snapshot as a space-saving copy of the source
Snapshot vs. Clone device, where capacity was consumed only for data changed after the snapshot time. Clones, on
the other hand referred to full copy of the source device. With VMAX3, TimeFinder SnapVX
snapshots are always space-efficient. When they are linked to host-addressable target devices,
the user can choose if to keep the target devices space-efficient, or perform full copy.
VMAX3 TimeFinder TimeFinder SnapVX is the latest development in TimeFinder local replications software, offering
SnapVX higher scale and wider feature set, yet maintaining the ability to emulate legacy behavior.
Storage volume LUN or unit of storage used by VPLEX presented by the back-end arrays.
VPLEX device Logical layer where protection scheme applied to an extent or group of extents used by VPLEX.
Virtual volume Unit of storage presented from VPLEX front-end ports to hosts.
Front-end port VPLEX director port connected to host initiators (acts as a target).
Back-end port VPLEX director port connected to storage arrays (acts as an initiator).
VPLEX Director The central processing and intelligence of the VPLEX solution. There are redundant (A and B)
directors in each VPLEX Engine.
VPLEX Engine Consists of two directors and is the unit of scale for the VPLEX solution.
VPLEX cluster A collection of VPLEX engines in one rack, using redundant, private Fibre Channel connections as
the cluster interconnect.
Storage Array An intelligent, highly available, scalable, hyper consolidated, multi-tiered block storage frame.
For example, EMC CX4, VNX, Symmetrix DMX4, VMAX, VMAX3, and XtremIO.
VPLEX Local: This solution is appropriate for customers who desire non-disruptive and fluid mobility,
high availability, and a single point of management for heterogeneous block storage systems within
a single data center.
VPLEX Metro: This solution is for customers who desire non-disruptive and fluid mobility, high
availability, and a single point of management for heterogeneous block storage systems within and
across data centers at synchronous distance. The VPLEX Metro offering also includes the unique
capability to remotely export virtual storage across datacenters without the need for physical storage
at the remote site.
ProtectPoint
EMC ProtectPoint
EMC ProtectPoint provides storage-integrated data protection that complements existing EMC data protection
and availability solutions and demonstrates the latest proof point of the protection storage architecture. The
EMC protection storage architecture is a blueprint for data protection transformation and investment protection
(FIGURE 3) that focuses on three key areas: data management services, data source integration, and protection
storage. This also illustrates the decoupling of the management of protection (data management services) from
protection storage and data source integration, which is critical for software-defined IT that requires the
decoupling of the management, control, and data planes.
Specifically, ProtectPoint addresses two aspects of data source integration - the integration with primary storage
and applications. ProtectPoint is an industry first solution that protects data by copying it directly from its
source (primary storage) to the protection storage via the most efficient path and without application impact. To
achieve this, ProtectPoint leverages key technologies within the primary storage and protection storage and
introduces new protection software. This protection software is a data protection agent that drives the backup
process and supports integration with the application being protected. This agent also enables the application
administrator to control his own backup and recovery operations.
ProtectPoint is neither adding application integration to snapshots nor adding snapshot support to backup
software. Both of these approaches would bring some benefits, but would not fully address the problems with
backup and would inevitably experience many of the limitations of snapshots. Rather, ProtectPoint was
designed by decoupling the data plane from the control plane to directly drive the underlying capabilities on the
primary and protection storage.
EMC ProtectPoint
The data plane carries the data from source to destination. With ProtectPoint, the data plane ( Figure 4) is the
connection between primary storage to the Data Domain system. Since ProtectPoint leverages primary storage
change block tracking technology, it minimizes data sent on the data plane. When a backup is triggered, unlike
a traditional backup application, the primary storage knows exactly what has changed since the last backup
and only has to send those unique blocks across the network. The direct data movement from primary to
protection storage eliminates the local area network impact by isolating all data traffic to the SAN. In addition,
unlike other backup mechanisms that consume valuable host side resources on the primary storage,
ProtectPoint data movement is handled by separate resources of the primary storage that are dedicated to
protection workflows.
ProtectPoint is very different from snap and replication solutions thanks to the efficient way the data is
processed and stored by the protection storage system. One of the benefits of leveraging Data Domain
protection storage is its industry leading inline deduplication technology. When the Data Domain system
While the data plane carries out the data movement and processing, the control plane coordinates each of the
steps along with other related activities.
EMC ProtectPoint for File Systems uses an agent with a simple command line interface to drive the ProtectPoint
flows with simple scripting to orchestrate with any application along with any additional orchestration that may
be needed such as additional local snapshots and in the case of this white paper VPLEX.
You can use ProtectPoint for File System to complete the following tasks:
Create a snapshot of the production application LUNs on the primary storage system.
Trigger the movement of data created from the backups on the primary storage system to the Data
Domain devices.
Create a static-image for each LUN in the data set on the Data Domain system.
Manage replication from the source Data Domain system to a Data Domain system in the data recovery
site.
Securely manage the credentials for the Data Domain systems.
Manage the ProtectPoint backup and restore catalog.
Manage the lifecycles of the data backups by listing and optionally deleting existing backups.
Show the ProtectPoint version number.
Validate the content and format of the configuration files.
The EMC VMAX3 family of storage arrays is built on the strategy of simple, intelligent, modular storage, and
incorporates a Dynamic Virtual Matrix interface that connects and shares resources across all VMAX3 engines,
allowing the storage array to seamlessly grow from an entry-level configuration into the worlds largest storage
array. It provides the highest levels of performance and availability featuring new hardware and software
capabilities.
The newest additions to the EMC VMAX3 family, VMAX 100K, 200K and 400K, deliver the latest in Tier-1 scale-
out multi-controller architecture with consolidation and efficiency for the enterprise. It offers dramatic
increases in floor tile density, high capacity flash and hard disk drives in dense enclosures for both 2.5" and
3.5" drives, and supports both block and file (eNAS).
VMAX3 family of storage arrays come pre-configured from factory to simplify deployment at customer sites and
minimize time to first I/O. Each array uses Virtual Provisioning to allow the user easy and quick storage
VMAX3 FAST.X
Previously named Federated Tiered Storage (FTS), FAST.X is a feature of VMAX3 that allows an external storage
to be connected to the VMAX3 backend and provide physical capacity that is managed by VMAX3 software.
Attaching external storage to a VMAX3 enables the use of physical disk capacity on a storage system that is
not a VMAX3 array, while gaining access to VMAX3 features, including cache optimizations, local and remote
replications, data management, and data migration.
The external storage devices can be encapsulated by VMAX3, and therefore their data preserved and
independent of VMAX3 specific structures, or presented as raw disks to VMAX3, where HYPERMAX OS will
initialize them and create native VMAX3 device structures.
FTS is implemented entirely within HYPERMAX OS and doesnt require any additional hardware besides the
VMAX3 and the external storage. Connectivity with the external array is established using fiber channel ports.
1
Fully Automated Storage Tiering (FAST) allows VMAX3 storage to automatically and dynamically manage performance service level goals across the
available storage resources to meet the application I/O demand, even as new data is added, and access patterns continue to change over time.
Data Domain systems reduce the amount of disk storage needed to retain and protect data by 10 to 30 times.
Data on disk is available online and onsite for longer retention periods, and restores become fast and reliable.
Storing only unique data on disk also means that data can be cost-effectively replicated over existing
networks to remote sites for DR. With the industrys fastest deduplication storage controller, Data Domain
systems allow more backups to complete faster while putting less pressure on limited backup windows.
All Data Domain systems are built as the data store of last resort, which is enabled by the EMC Data Domain
Data Invulnerability Architecture end-to-end data verification, continuous fault detection and self-healing,
and other resiliency features transparent to the application.
For more information on Data Domain visit: http://www.emc.com/data-protection/data-domain/index.htm
Table 1 depicts the basic Data Domain vdisk block device object hierarchy, which is also shown in Figure 8.
Name Description
Pool Similar to a Department level. Maximum of 32 pools.
Device Group Similar to the Application level. Maximum of 1024 device groups per pool.
Note: Backup and restore operations of backup-sets are performed at a Data Domain device group
granularity. Therefore, when preparing Data Domain devices, it is critical that all the Data Domain
devices that participate in the same backup-set will belong to the same Data Domain device group.
Remember that there will be two identical sets of devices: Data Domain backup devices, and Data
Domain restore devices, and they all have to be part of the same device group.
Note: Data Domain can replicate the backups to another Data Domain system by using Data Domain
Replicator (separately licensed). While Data Domain file system structure is not covered in this paper,
the reader should be aware that the replication granularity is either single backup-set, or a vdisk Pool2.
2
This feature is not currently available with Data Domain OS 5.5. Refer to Data Domain OS release notes for available
features.
App
OS
Production Production Point In Unique Backup
Data Data Time Copy Blocks Images
Ingested
SAN Inline
CBT Transfer
SAN Dedupe
Virtual
Change Block Tracking Deduplication Index
Storage
SAN Raid-1
Inline
CBT Transfer
SAN Dedupe
Virtual
Change Block Tracking Deduplication Index
Storage
In Figure 10, the VPLEX virtual volume (indicated in orange) sits on top of a RAID-1 device (indicated in green).
Each RAID-1 device has two mirror legs based on a single extent (indicated in blue and dark blue). Each of the
two extents is equal in size to the underlying storage volume from the VMAX3 and from a second (possibly non-
VMAX3) array. The key difference in this topology is the mirror legs -- only one of them will be backed up using
ProtectPoint. Though it is possible both legs could be from the same VMAX3 or two different VMAX3 arrays, the
best practice configuration would only leverage ProtectPoint on one mirror leg.
Note: It is also possible to design complex multi-site backup solutions using VPLEX
distributed raid-1 devices and VMAX3 storage at two data centers. These solutions would
provide local backup and restore capabilities at each data center without the need to
replicate the backups from one DD device to the other. For restore, however, only one site
can do restore to a distributed device and a full rebuild would be required for the non-
restored mirror leg (as shown later in this document).
Note: Applications that span multiple VMAX3 arrays are able to use SnapVX to achieve
consistent snapshots, but currently ProtectPoint does not support this configuration.
Regardless of the chosen RAID-1 configuration, the determination of which mirror leg is being used for
ProtectPoint backup becomes pivotal for a successful backup (and restore). The objective remains the same
as with Figure 9 - to create a backup of a VPLEX virtual volume based on a 1:1 mapped VMAX3 extent. The
required checks to determine the VPLEX volume device geometry, mirror legs, and mirror leg heath are
provided in following procedure sections.
Data Domain WWN of the Copies the static-image that contains the backup image from
encapsulated restore target devices the primary Data Domain system to the secondary Data
Domain system.
Block services for ProtectPoint pool
and device-group on the secondary
Data Domain system
Primary Data Domain system Information for the primary Data Domain system. Used for all
hostname or IP address, username, control path operations for the Data Domain system.
and password
Secondary Data Domain system Information for the secondary Data Domain system. Used for
hostname or IP address, username, static-image or MTree replication.
and password.
From the GUI navigate to the virtual volumes context, select (mouse-over) the virtual volume(s)
and then click the View Map hyperlink:
The underlying VMAX3 volume(s) will be shown for each of the above methods. Use the VMAX3 VPD ID
or the VMAX3 Storage Volume Name to determine the VMAX3 Symdev-ID within the VMAX3. The
VMAX3 Symdev-ID can be determined using information contained within the storage-volumes context
in the VPLEX CLI or GUI:
In this case the virtual volume mapping output is used to determine the VMAX3 Symdev-ID. The
configuration file for ProtectPoint 2.0 is modified to include the VMAX3 information in order to make
SnapVX snapshots of the VMAX3 volume using the FTS mapped DataDomain vdisk.
Note: Each of the following checks are typically scripted or built into code that orchestrates
the overall backup process.
a. Confirm a VPLEX code upgrade is not in progress. Using the VPLEX CLI issue the ndu status
command and confirm that the response is No firmware, BIOS/POST, or SSD upgrade is in
progress.
b. Confirm the device is healthy. Issue the ll command from the /clusters/<cluster name>/virtual-
volumes/<virtual volume name> context for each volume(s) to be copied.
i. Confirm the underlying device status is not marked out of date or in a rebuilding state.
ii. Confirm Health Status is ok
iii. Confirm Operational Status is ok
c. Confirm the source VPLEX virtual volume device geometry is not RAID-C. Device geometry can be
determined by issuing the ll command at the /clusters/<cluster name>/devices/<device name>
context.
d. Confirm each volume is 1:1 mapped (single extent) RAID-0 or 1:1 mapped (two extent) local RAID-
1. Distributed RAID-1 device legs must be a combination of RAID-0 (single extent) and/or RAID-1
(two extents) device geometries.
e. Confirm the device is not being protected by RecoverPoint (Note: ProtectPoint restore to a
RecoverPoint protected device would lead to data inconsistency). Issue the ll command from
the /clusters/<cluster name>/virtual-volumes/<virtual volume name> context and check
recoverpoint-protection-at is set to [] and recoverpoint-usage is set to -.
f. Confirm VPLEX volumes to be copied do not have remote locality (from same VPLEX cluster).
Issue the ll command against the /clusters/<local cluster name>/virtual volumes/<virtual
volume name> context and confirm locality is local or distributed. The goal here is to ensure
the host, VMAX3 and DD are all in the same data center.
g. Ensure virtual volumes are members of the same VPLEX consistency group. In most cases all
members of the consistency should be backed up together. Consistency group membership can
be determined by issuing ll from the /clusters/<cluster name>/consistency-groups/<consistency
group name> context.
Note: VPLEX consistency group membership should align with any VMAX3 storage group
membership whenever possible.
h. For RAID-1 or distributed RAID-1 based virtual volumes, confirm underlying VMAX3 storage
volume status is not failed or in an error state. Issue the ll command from the
/clusters/<cluster name>/devices context or from /distributed-storage/distributed-
devices/<distributed device name>/components context.
Note: Best practice is to set the VPLEX consistency group detach rule (winning site) to
match site where VMAX3 backups are performed.
3. If any of the above conditions fail, the backup should be aborted. If the conditions are being checked by
a script an error message or alert should be generated.
4. Follow standard ProtectPoint for File Systems procedures to create a backup of the host file system(s).
The ProtectPoint File System Agent 2.0 Installation and Administration guide located at
https://support.emc.com contains the detailed steps.
Using ProtectPoint Recovery Device Images for Granular Recovery with VPLEX
Once a backup is created it can be presented (zoned and lun masked) directly from the VMAX3 back to a
test or development host. In some situations the administrator may wish to present the backup image
through VPLEX to the production host to perform a granular restore. Figure 11 below shows the granular
restore process using ProtectPoint for File Systems with VMAX3 and VPLEX.
App
OS
Production Production Point In Unique Backup
Data Data Time Blocks Images
Copy Ingested
SAN
SAN Recovery
Recovery
Device Device
Granular Recovery
Mapping Dedupe
Index
For ProtectPoint Recovery Device images presented back through VPLEX, perform these additional VPLEX
specific steps:
1. Confirm the VMAX3 Recovery Device images are visible to VPLEX BE ports. As necessary, perform
VMAX3 to VPLEX masking/storage group modification to add the backup images. See the
ProtectPoint 2.0 administration guide available on https://support.emc.com for details on selecting
the correct Recovery Image (backup) and making it available through VMAX3.
Mapping Dedupe
Virtual
Index
Storage
or
cache-invalidate-status
-----------------------
director-1-1-A status: in-progress
result: -
cause: -
cache-invalidate-status
-----------------------
director-1-1-A status: completed
result: successful
cause: -
Note: For pre-5.2 code, it is recommended to wait a minimum of 30 seconds to ensure that
the VPLEX read cache has been invalidated for each virtual volume. This can be done
concurrently with Step 3. With GeoSynchrony 5.3 P3 and higher there is no longer a need to
wait 30s. The cache-invalidate command now runs as a foreground process and not as an
asynchronous background request.
6. Follow standard ProtectPoint for File Systems procedures to perform a full restore the host file
system(s). The ProtectPoint File System Agent 2.0 Installation and Administration guide located
at https://support.emc.com contains the detailed full restore steps.
7. Confirm the IO Status of the production VMAX3 storage volumes within VPLEX is alive by doing a long
listing against the storage volumes context for your cluster.
For example:
In addition, confirm VPLEX back-end paths are healthy by issuing the connectivity validate-be
command from the VPLEX CLI. Ensure that there are no errors or connectivity issues to the back-end
storage devices. Resolve any error conditions with the back-end storage before proceeding.
Example output showing desired back-end status:
8. For Pre 5.2 code, restore access to virtual volume(s) based on source devices for host(s): Add the virtual
volume back to the view, specifying the original LUN number (noted in step 2) using VPLEX CLI:
/clusters/<cluster name>/exports/storage-views> addvirtualvolume -v
storage_view_name/ -o (lun#, virtual_volume_name) -f
9. As necessary, rescan devices and restore paths (for example, powermt restore) on hosts.
10. As necessary, mount devices.
App
OS
Production Production Point In Unique Backup
Data Data Time Blocks Images
Copy Ingested
SAN
Raid-1
Mapping Dedupe
Virtual
Index
Storage
Figure 15: ProtectPoint 2.0 Full Restore of VPLEX RAID-1 Virtual Volumes
Technical Note: This same set of steps can be applied to remote array-based copy
products like SRDF or MirrorView. For example, an SRDF R2 or MirrorView Secondary Image
is essentially identical in function to local array-based copy. The remote copy, in this case,
can be used to do a restore to a production (R1/Primary Image) volume.
Prerequisites
This section assumes VPLEX virtual volumes are based on distributed or local RAID-1 VPLEX virtual volumes
built from at least one VMAX3 array with ProtectPoint. In addition, the VPLEX virtual volumes must possess
both of the following attributes:
Note: Each of the following checks are typically scripted or built into code that orchestrates the
overall restore process on the array.
a. Confirm a VPLEX ndu is not in progress. Using the VPLEX CLI issue the ndu status command
and confirm that the response is No firmware, BIOS/POST, or SSD upgrade is in progress.
b. Confirm the restore target device is healthy. Issue the ll command from the
/clusters/<cluster name>/virtual-volumes/<virtual volume name> context for each volume(s)
to be copied.
i. Confirm the underlying device status is not marked out of date or in a rebuilding
state.
ii. Confirm Health Status is ok
iii. Confirm Operational Status is ok
c. Confirm the underlying VPLEX device geometry is not RAID-c. Device geometry can be
determined by issuing ll at the /clusters/<cluster name>/devices/<device name> context.
d. Confirm each volume is 1:1 mapped (single extent) RAID-0 or 1:1 mapped (two extent) local
RAID-1. Distributed RAID-1 device legs must be a combination of RAID-0 (single extent)
and/or RAID-1 (two extent) device geometries.
e. Confirm the VMAX3 production device is not being protected by RecoverPoint. Issue the ll
command from the /clusters/<cluster name>/virtual-volumes/<virtual volume name> context
and check recoverpoint-protection-at is set to [] and recoverpoint-usage is set to -.
f. Confirm VPLEX volumes being restored have the same locality (from same VPLEX cluster) and
are in the same array.
g. Ensure virtual volumes are members of the same VPLEX consistency group. In most cases all
members of the consistency should be copied together. Consistency group membership can
be determined by issuing ll from the /clusters/<cluster name>/consistency-
groups/<consistency group name> context.
h. Confirm underlying VMAX3 storage volume status is not failed or in an error state. Issue the
ll command from the /clusters/<cluster name>/devices context or from /distributed-
storage/distributed-devices/<distributed device name>/components context.
Note: Best practice is to set the VPLEX consistency group detach rule (winning site) to
match site where the VMAX3 and Data Domain systems reside.
2. Shut down any host applications using the production VPLEX virtual volume(s) or file system that will be
restored. As necessary, unmount the associated virtual volumes on the host. The objectives here are
to prevent host access and to clear the hosts read cache. Depending on how things have been
scripted, ProtectPoint for File System will handle taking the filesystem offline / unmounting prior to
restore.
3. Invalidate VPLEX read cache on the virtual volume(s) being restored. There are several options to
achieve VPLEX read cache invalidation depending on your VPLEX GeoSynchrony code version:
A. For pre-5.2 code, remove the source virtual volume(s) from all storage views. Make note of the
virtual volume lun numbers within the storage view prior to removing them. You will need this
information in step 7 below.
or
B. For 5.2 and higher code,
1. Use virtual-volume cache-invalidate to invalidate the individual source volume(s):
VPlexcli:/> virtual-volume cache-invalidate <virtual volume>
or
Use consistency-group cache-invalidate to invalidate an entire VPLEX consistency group of
source volumes.
VPlexcli:/> consistency-group cache-invalidate <consistency group>
cache-invalidate-status
-----------------------
director-1-1-A status: in-progress
result: -
cause: -
cache-invalidate-status
-----------------------
director-1-1-A status: completed
result: successful
cause: -
Note: For pre-5.2 code, it is recommended to wait a minimum of 30 seconds to ensure that
the VPLEX read cache has been invalidated for each virtual volume. This can be done
concurrently with Step 3.
Note: Cache-invalidate command must not be executed on the Recover Point enabled virtual-
volumes. This means using either RecoverPoint or array-based copies, but not both, for a
given virtual volume. The VPLEX Clusters should not be undergoing a NDU while this
command is being executed.
4. Detach the VPLEX device RAID-1 or Distributed RAID-1 mirror leg that will not be restored during the array-
based restore processes. This will typically be the non-VMAX3 mirror leg. If the virtual volume is a
member of a consistency group, in some cases the virtual volume may no longer have storage at one
site which may cause the detach command to fail. In this case the virtual volume will need to be
removed from the consistency group *before* it the mirror leg is detached. Use the detach-mirror
command to detach the mirror leg(s):
device detach-mirror -m <device_mirror_to_detach> -d <distributed_device_name> i f
Note: Depending on the raid geometry for each leg of the distributed device, it may be
necessary to detach both the local mirror leg and the remote mirror leg. This is because
only 1 storage volume is being restored and there are up to 3 additional mirrored copies
maintained by VPLEX (1 local and 1 or 2 remote). For example, if the VPLEX distributed
device mirror leg being restored is, itself, a RAID-1 device then both the non-restored local
leg and the remote leg must be detached.
6. Confirm the IO Status of storage volumes based on array-based clones is alive by doing a long listing
against the storage volumes context for your cluster.
For example:
In addition, confirm VPLEX back-end paths are healthy by issuing the connectivity validate-be command
from the VPLEX CLI. Ensure that there are no errors or connectivity issues to the back-end storage devices.
Resolve any error conditions with the back-end storage before proceeding.
Example output showing desired back-end status:
Distributed RAID-1:
device attach-mirror -m <2nd mirror leg to attach> -d /clusters/<cluster
name>/devices/<existing distributed RAID-1 device>
8. For Pre 5.2 code, restore host access to the VPLEX volume(s).
If the virtual volume is built from a local RAID 1 device:
The lun# is the previously recorded value from step 2 for each virtual volume.
Note: EMC recommends waiting at least 30 seconds after removing access from a storage
view to restore access. This is done to ensure that the VPLEX cache has been cleared for
the volumes. The array-based restore will likely will take 30 seconds, but if you are
scripting be sure to add a pause.
Note: Some hosts and applications are sensitive to LUN numbering changes. Use the
information you recorded in step 3 to ensure the same LUN numbering when you restore
the virtual volume access.
Note: Full mirror synchronization is not required prior to restoring access to virtual
volumes. VPLEX will synchronize the second mirror leg in the background while using the
first mirror leg as necessary to service reads to any unsynchronized blocks.
References
The following reference documents are available at Support.EMC.com:
White Paper: Workload Resiliency with EMC VPLEX
EMC ProtectPoint File System Agent with VMAX3 Backup & Recovery Best Practices for Oracle on
ASM