Vous êtes sur la page 1sur 172

Using Celerra Replicator (V1)

P/N 300-004-184
Rev A08

Version 5.6.47
December 2009

Contents
Introduction to Celerra Replicator . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .5
Terminology . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .5
Restrictions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .7
Cautions. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .9
Celerra Replicator concepts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .10
Local replication . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .10
Remote replication . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .11
Activating the destination file system as read/write . . . . . . . . . . . . .13
Communication between Celerra Network Servers. . . . . . . . . . . . . .17
How resynchronization works . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .18
How suspend works . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .19
How replication relationship restarts . . . . . . . . . . . . . . . . . . . . . . . . .20
System requirements for Celerra Replicator. . . . . . . . . . . . . . . . . . . . . . .22
Local replication . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .22
Remote replication . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .22
Upgrading from previous Celerra Network Server versions . . . . . . . . . .24
Upgrade from a version earlier than 5.5.39.2 . . . . . . . . . . . . . . . . . . .24
Upgrade from Celerra Network Server version 5.5.39.2 or later. . . .25
Planning considerations for Celerra Replicator . . . . . . . . . . . . . . . . . . . .29
Replication policies . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .29
SavVol size requirements for remote replication . . . . . . . . . . . . . . .33
Determine the number of replications per Data Mover . . . . . . . . . . .34
Configuration considerations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .35
User interface choices for Celerra Replicator. . . . . . . . . . . . . . . . . . . . . .37
Roadmap for Celerra Replicator . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .38
Initiating replication. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .39
Task 1: Establish communication . . . . . . . . . . . . . . . . . . . . . . . . . . . .40
Task 2: Verify communication . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .40
Task 3: Create SnapSure checkpoint of source file system. . . . . . .42
Task 4: Create the destination file system . . . . . . . . . . . . . . . . . . . . .45
Task 5: Copy checkpoint to the destination file system . . . . . . . . . .45
Task 6: Begin replication . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .48
Task 7: Create a second checkpoint of the source file system . . . .50
Task 8: Copy incremental changes. . . . . . . . . . . . . . . . . . . . . . . . . . .52

1 of 172
Task 9: Verify file system conversion . . . . . . . . . . . . . . . . . . . . . . . . 54
Task 10: Check replication status . . . . . . . . . . . . . . . . . . . . . . . . . . . 55
Task 11: Create restartable checkpoints . . . . . . . . . . . . . . . . . . . . . . 59
Recovering replication data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 61
Task 1: Replication failover. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 62
Task 2: Resynchronize the source and destination sites . . . . . . . . 66
Task 3: Replication reversal . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 74
Abort Celerra Replicator. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 79
Suspend a replication relationship . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 81
Verify the suspended replication relationship . . . . . . . . . . . . . . . . . 88
Restarting a replication relationship . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 89
Verify that the replication relationship is not synchronized . . . . . . 90
Restart replication relationship . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 91
Extending the size of a file system . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 98
Extend file system size automatically . . . . . . . . . . . . . . . . . . . . . . . . 98
Extend file system size . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 101
Resetting replication policy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 105
High water mark and time-out policies . . . . . . . . . . . . . . . . . . . . . . 105
Modify replication policy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 106
Change flow-control policies . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 107
Set bandwidth size . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 109
Set policies using parameters . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 110
Reverse the direction of a replication relationship . . . . . . . . . . . . . . . . 111
Verify the reverse direction of replication relationship . . . . . . . . . 113
Monitor replication . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 114
Checking playback service and outstanding delta sets . . . . . . . . . . . . 115
Task 1: Determine playback service status . . . . . . . . . . . . . . . . . . . 115
Task 2: Playback delta set. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 118
Task 3: Verify delta set . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 119
Events for Celerra Replicator. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 121
Change the Celerra Replicator SavVol default size . . . . . . . . . . . . . . . . 123
Change the passphrase between Celerra Network Servers . . . . . . . . . 124
Managing and avoiding IP replication problems . . . . . . . . . . . . . . . . . . 125
Preventive measures to avoid IP replication problems . . . . . . . . . 125
Replication restart methods . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 128
Recovering from a corrupted file system . . . . . . . . . . . . . . . . . . . . 130
Managing anticipated destination site or network outages . . . . . . 131
Managing unanticipated destination site or network outages. . . . 132
Managing unanticipated source site outages . . . . . . . . . . . . . . . . . 133
Managing expected source site outages . . . . . . . . . . . . . . . . . . . . . 133
Mount the destination file system read/write temporarily . . . . . . . 133
Recovering from an inactive replication state. . . . . . . . . . . . . . . . . 135
Creating checkpoints on the destination site . . . . . . . . . . . . . . . . . 136
Copy file system to multiple destinations with fs_copy. . . . . . . . . 136
Transporting replication data using disk or tape . . . . . . . . . . . . . . . . . . 139
Disk transport method . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 140
Tape transport method . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 144
Setting up the CLARiiON disk array . . . . . . . . . . . . . . . . . . . . . . . . . . . . 147
Review the prerequisites . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 147
Run the setup script . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 149
Create data LUNs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 151
Troubleshooting Celerra Replicator . . . . . . . . . . . . . . . . . . . . . . . . . . . . 154
Where to get help . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 154

2 of 172 Version 5.6.47 Using Celerra Replicator (V1)


E-Lab Interoperability Navigator . . . . . . . . . . . . . . . . . . . . . . . . . . . .154
Log files for troubleshooting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .154
server_log messages . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .155
Network performance troubleshooting. . . . . . . . . . . . . . . . . . . . . . .156
Failure during transport of delta set . . . . . . . . . . . . . . . . . . . . . . . . .156
Failure of fs_copy command process . . . . . . . . . . . . . . . . . . . . . . .156
Control Station restarts during replication . . . . . . . . . . . . . . . . . . .156
Control Station fails over . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .157
NS series loses power . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .157
Return codes for fs_copy. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .157
Error messages for Celerra Replicator . . . . . . . . . . . . . . . . . . . . . . . . . .161
Related information . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .162
Training and Professional Services . . . . . . . . . . . . . . . . . . . . . . . . .163
Appendix A: fs_replicate -info output fields . . . . . . . . . . . . . . . . . . . . . .164
Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .169

Using Celerra Replicator (V1) Version 5.6.47 3 of 172


4 of 172 Version 5.6.47 Using Celerra Replicator (V1)
Introduction to Celerra Replicator
EMC® Celerra Replicator™ produces a read-only, point-in-time copy of a source file
system and periodically updates this copy, making it consistent with the source file
system. This read-only copy can be used by a Data Mover in the same Celerra
cabinet, or a Data Mover at a remote site for content distribution, backup, and
application testing.
This technical module is part of the EMC Celerra® Network Server documentation
set and is intended for system administrators who are responsible for IP replication
in the Celerra environment. Before using Celerra Replicator, system administrators
establishing replication should understand Celerra volumes and file systems.
This technical module is one of several technical modules that describe replication
using different implementations of Celerra Replicator. Use the following guidelines
to navigate the technical modules:
◆ Read this technical module to learn the basics about the Celerra Replicator
product and how it performs local replication or remote replication of Production
File Systems (PFSs) over an IP network to a destination. This technical module
describes how to set up the IP environment for replication and how to use
Celerra Replicator to replicate your file systems. Because Celerra Replicator
relies on EMC SnapSure™ checkpoints (read-only, logical, point-in-time
images) for the initial copy and for replication restart, you might find it useful to
read Using SnapSure on EMC Celerra to learn more about checkpoints.
◆ Read Replicating EMC Celerra CIFS Environments (V1) to learn how to use
Celerra Replicator to perform replication in the CIFS environment. Replicating
EMC Celerra CIFS Environments (V1) describes how to replicate the CIFS
environment information in the root file system of a Virtual Data Mover (VDM) as
well as how to replicate the file systems mounted to that VDM. To learn about
VDMs in general, read Configuring Virtual Data Movers for EMC Celerra.
◆ Read Using EMC Celerra Replicator for iSCSI (V1) to learn about iSCSI
replication using Celerra Replicator for iSCSI product. Using EMC Celerra
Replicator for iSCSI (V1) describes how to replicate production iSCSI LUNs by
asynchronously distributing local point-in-time copies of the LUNs to a
destination.
◆ Read Using EMC Celerra Replicator (V2) to learn about the new Celerra
Replicator product.
◆ Read Managing EMC Celerra Volumes and File Systems with Automatic
Volume Management and Managing EMC Celerra Volumes and File Systems
Manually to learn more about establishing replication.

Terminology
This section defines terms important to understanding replication on the Celerra
Network Server. The EMC Celerra Glossary provides a complete list of Celerra
terms.
Automatic File System Extension: A configurable Celerra file system feature that
automatically extends a file system created or extended with Automatic Volume

Using Celerra Replicator (V1) Version 5.6.47 5 of 172


Manager (AVM) when the high water mark (HWM) is reached. See also high water
mark.
checkpoint: Read-only, logical, point-in-time image of a file system. A checkpoint is
sometimes referred to as a checkpoint file system or a SnapSure file system.
delta set: Set containing the block modifications made to the source file system that
Celerra Replicator uses to update the destination file system (read-only, point-in-
time, consistent replica of the source file system). The minimum delta-set size is
128 MB.
high water mark (HWM): The trigger point at which Celerra Network Server performs
one or more actions, such as sending a warning message, extending a volume, or
updating a replicated file system, as directed by the related feature’s
software/parameter settings.
IP replication service: The service that uses the IP network to transfer the delta sets
from the replication SavVol on the source site to the replication SavVol on the
destination site.
local replication: Replication of a file system on a single Celerra Network Server
with the source file system on one Data Mover and the destination file system on
another Data Mover.
loopback replication: Replication of a file system with the source and destination file
systems residing on the same Data Mover.
playback service: The process of reading the delta sets from the destination SavVol
and updating the destination file system.
remote replication: Replication of a file system from one Celerra Network Server to
another. The source file system resides on a different Celerra system from the
destination file system.
replication: A service that produces a read-only, point-in-time copy of a source file
system. The service periodically updates the copy, making it consistent with the
source file system.
replication reversal: The process of reversing the direction of replication. The
source file system becomes read-only and the destination file system becomes
read/write.
replication service: Service that copies modified blocks from the source file system
to a replication SavVol prior to transferring the data to the destination file system.
Replicator ConfigVol: An internal information store for replication. Provides a
storage vehicle for tracking changes in the source file system.
Replicator failover: The process that changes the destination file system from read-
only to read/write and stops the transmission of replicated data. The source file
system, if available, becomes read-only.
Replicator SavVol: A Celerra volume, required by replication, used to store modified
data blocks from the source file system.
SnapSure SavVol: A Celerra volume to which SnapSure copies point-in-time data
blocks from the PFS before the blocks are altered by a transaction. SnapSure uses
the contents of the SavVol and the unchanged PFS blocks to maintain a checkpoint
of the PFS.
timeout: Time interval at which the system takes a predetermined action.

6 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Virtual Data Mover (VDM): A Celerra software feature enabling users to
administratively separate CIFS servers, replicate CIFS environments, and move
CIFS servers from Data Mover to Data Mover.
Virtual Provisioning: A configurable Celerra file system feature that can only be
used in conjunction with Automatic File System Extension. This option lets you
allocate storage based on longer-term projections, while you dedicate only the file
system resources you currently need. Users—NFS or CIFS clients and application
—see the virtual maximum size of the file system of which only a portion is
physically allocated. Combined, the Automatic File System Extension and Virtual
Provisioning options let you grow the file system gradually on an as-needed basis.

Restrictions
The following restrictions apply to Celerra Replicator:
◆ Celerra Data Migration Service (CDMS) is unsupported (an mgfs file system
cannot be replicated).
◆ Multi-Path File System (MPFS) is supported on the source file system, but not
on the destination file system.
◆ EMC E-Lab™ Interoperability Navigator provides information about disaster
recovery replication products such as EMC SRDF®/Synchronous (SRDF/S) and
SRDF/Asynchronous (SRDF/A).
◆ For EMC TimeFinder®/FS:
• A business continuance volume (BCV) cannot be a source or a destination
file system for replication. You can replicate the underlying source file
system, but not the BCV.
• Do not use the TimeFinder/FS -Restore option for a replicated source file
system. Replication will be unaware of any changes because these changes
occur at the volume level. However, you can restore on a single file basis
using a NFS/CIFS client, which has access to the source file system and the
BCV of the source file system.
• Do not use TimeFinder/FS with a file system that was created on a slice
volume. Creating a file system using the samesize option slices the volume.
TimeFinder does not recognize sliced partitions. Using TimeFinder/FS,
NearCopy, and FarCopy with EMC Celerra further details this feature.
◆ For TimeFinder/FS Near Copy and Far Copy:
• A BCV cannot be a source or a destination file system for replication. You
can replicate the underlying source file system, but cannot replicate the BCV.
◆ Do not extend the source file system while fs_copy is running.
◆ On a per-Data-Mover basis, the total size of all file systems, the size of all
SavVols used by SnapSure, and the size of all SavVols used by the Celerra
Replicator feature must be less than the total supported capacity of the Data
Mover. The EMC Celerra Network Server Release Notes, available at
http://Powerlink.EMC.com, the EMC Powerlink® website, provide a list of Data
Mover capacities.

Using Celerra Replicator (V1) Version 5.6.47 7 of 172


◆ When replicating databases, additional application-specific actions may be
necessary to bring the database to a consistent state (for example, quiescing
the database).
◆ If you plan to enable international character sets (Unicode) on your source and
destination sites, you must first set up translation files on both sites before
starting Unicode conversion on the source site. Using International Character
Sets with EMC Celerra describes this action in detail.
◆ In the case of multiple file systems, all fs_replicate commands must be
executed sequentially.
◆ IP replication failover is not supported for local groups unless you use VDMs.
Configuring Virtual Data Movers for EMC Celerra describes VDM configuration.
◆ IP replications created prior to Celerra Network Server version 5.5 in which the
source file system contained iSCSI LUNs are no longer supported. These
replications will continue to run in their current state, but you cannot actively
manage (suspend or resume) them. Any attempt to perform such an operation
prompts an error stating that the “item is currently in use by iSCSI.” You can
abort or delete the replication. EMC recommends that you convert any existing
IP replications to iSCSI replications as soon as possible.
◆ Do not use IP aliasing for IP replication. Use IP aliasing only for Control Station
access from the client.
◆ For full management capability, ensure that you have the same Celerra version
on the source and destination Celerra Network Servers. For example, you
cannot have version 5.5 on the source side and version 5.6 on the destination
side.
For limited management capability and no fs_copy support, out-of-family
replication support is available and requires NAS version 5.5.39.2 or later on
the source Celerra and 5.6.47 or later on the destination Celerra. "Upgrade from
Celerra Network Server version 5.5.39.2 or later" on page 25 provides more
details.
◆ For FLR-E-enabled file systems, both the source and destination file systems
must have the same FLR setting enabled. For example, if the source file system
has FLR-E enabled, then the destination file system must also have FLR-E
enabled. If the source file system does not have FLR-E enabled, then the
destination cannot have FLR-E enabled.
◆ File systems enabled for processing by Celerra Data Deduplication cannot be
replicated using Celerra Replicator. In addition, you cannot enable
deduplication on a file system that is already being replicated by Celerra
Replicator.
◆ In Celerra Replicator (V1) version 5.6, the fs_replicate -refresh command does
not create a new delta set until the production file system is updated.
Consequently, policy values for timeout, high water mark (HWM), autofreeze,
autoreadonly, and so on, that are modified using the fs_replicate -modify or
fs_replicate -refresh command, remain ineffective until a new delta set is
created.
In versions prior to 5.6, the fs_replicate -refresh command creates a new delta
set even if no updates are made to the PFS.

8 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Cautions
This section lists the cautions for using this feature on Celerra Network Server. If
any of this data is unclear, contact EMC Customer Service for assistance:
◆ To provide a graceful shutdown in an electrical power loss, Celerra Network
Server and the storage array need to have Uninterruptible Power Supply (UPS)
protection. If this is not provided, replication will become inactive and might
result in data loss. If replication becomes inactive, consult "Restarting a
replication relationship" on page 89 to determine if you can resume the
replication relationship.
◆ Replicating file systems from a Unicode-enabled Data Mover to an ASCII-
enabled Data Mover is not supported. I18N mode (Unicode or ASCII) must be
the same on the source and destination Data Movers.
◆ Replication sessions should run serially, not concurrently. That is, they should
start one after the other, not simultaneously.

Using Celerra Replicator (V1) Version 5.6.47 9 of 172


Celerra Replicator concepts
This section explains how replication produces and uses a read-only, point-in-time
copy of a source file system on the same or different Celerra Network Servers. It
also describes how Celerra Replicator enables you to activate failover to the
destination site for production, if the source site experiences a disaster and is
unavailable for data processing.

Local replication
Replication produces a read-only copy of the source file system for use by a Data
Mover in the same Celerra cabinet. The source and destination file systems are
stored on separate volumes. Local replication can use different Data Movers or the
same Data Mover.
Local replication process
Figure 1 on page 10 and subsequent steps show the processing that occurs when
using local replication for the first time.

Celerra Network Server

Primary Secondary
Data Mover Data Mover

2
Source Destination
file file
system system

3 4

SavVol

Storage unit

CNS-000765

Figure 1 Local replication

10 of 172 Version 5.6.47 Using Celerra Replicator (V1)


The process is as follows:
1. Throughout this process, network clients read and write to the source file
systems through the primary Data Mover without interruption.
2. For the initial replication start, the source and destination file systems are
manually synchronized using the fs_copy command.
3. After synchronization, the replication service uses the addresses of all block
modifications made to the source file system to create one or more delta sets.
The modified blocks are copied to the SavVol shared by the primary and
secondary Data Movers.
4. The local replication playback service periodically reads any available,
complete delta sets and updates the destination file system, making it
consistent with the source file system. During this time, the system tracks all
subsequent changes made to the source file system.
5. The secondary Data Mover exports the read-only copy to use for content
distribution, backup, and application testing. This optional step is done
manually.

Remote replication
Remote replication creates and periodically updates a read-only copy of a source
file system at a remote (destination) site. This is done by transferring changes
made to a source file system at a local site to a file system replica (destination) at
the destination site over an IP network. These transfers are automatic and are
based on user-definable replication policies.

Using Celerra Replicator (V1) Version 5.6.47 11 of 172


Remote replication process
Figure 2 on page 12 and subsequent steps show the processing that occurs when
using remote replication for the first time:
Source site Destination site
Celerra Network Server 1 Celerra Network Server 2

Data Data
Mover Mover

2
Source Destination
file file
system system
3 5

4
Source Destination
SavVol SavVol

Source storage unit Destination storage unit

CNS-000762

Figure 2 Remote replication


1. Throughout this process, network clients read and write to the source file
systems through a Data Mover at the source site without interruption.
2. For the initial replication start, the source and destination file systems are
synchronized using the fs_copy command. This can be performed over the IP
network or, if the source file system contains a large amount of data, by
physically transporting the data to a remote site using disk or tape.
3. The addresses of all subsequent block modifications made to the source file
system are used by replication to create one or more delta sets. The replication
service creates a delta set by copying the modified blocks to the SavVol at the
source site.
4. Replication transfers any available, complete delta sets which include the block
addresses to the destination SavVol. During this time, the system tracks
subsequent changes made to the source file system on the source site.
5. At the destination site, the playback service plays back any available, complete
delta sets to the destination file system, which makes it consistent with the
source file system.

12 of 172 Version 5.6.47 Using Celerra Replicator (V1)


6. The Data Mover at the destination site exports the read-only copy for content
distribution, backup, and application testing. This optional step is done
manually.

Activating the destination file system as read/write


If the source file system becomes unavailable, usually as the result of a disaster,
you can make the destination file system read/write for local or remote scenarios.
After the source site is available again, you can then restore replication to become
read/write at the source site and read-only at the destination site. This three-stage
process includes:
◆ Using the destination file system for production when the source file system is
unavailable (failover).
◆ Resynchronizing the file systems.
◆ Restoring the replication process to its original state.
Replication failover
In this example, the source site has experienced a disaster and is unavailable.
Failover ends the replication relationship between the source and destination file
systems and changes the destination file system from read-only to read/write.
When failing over, the following actions occur:
1. The system stops replication and plays back the outstanding delta sets on the
destination site to the destination file system according to the options specified
by the user. The system can play back either all or none of the delta sets. The
beginning of the failover process is shown in Figure 3 on page 14.

Using Celerra Replicator (V1) Version 5.6.47 13 of 172


Source site Destination site
Celerra Network Server 1 Celerra Network Server 2

Data Data
Mover Mover

Source Destination
file file
system system

Source Destination
SavVol SavVol

Source storage unit Destination storage unit

CNS-000763

Figure 3 Source site becomes unavailable


2. The system stops the playback service and the destination file system becomes
read/write, as shown in Figure 4 on page 15. This illustration displays a remote
replication scenario with the destination site as read/write. This state is possible
in a local replication scenario as well.

14 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Source site Destination site
Celerra Network Server 1 Celerra Network Server 2

Data Data
Mover Mover

Source Destination
file file
system system

Source Destination
SavVol SavVol

Source storage unit Destination storage unit

CNS-000764

Figure 4 Failover
3. The destination site can be enabled to allow read/write access to the
destination file system from network clients (local or remote scenario). "After a
failover or reversal" on page 78 provides more information. This optional step is
done manually.

Note: If the source file system is online, it becomes read-only.

Using Celerra Replicator (V1) Version 5.6.47 15 of 172


Resynchronization
When the original source file system becomes available, the replication relationship
can be reestablished. The fs_replicate -resync option is used to populate the
source file system with the changes made to the destination file system while the
source site was unavailable. This establishes the replication relationship in the
reverse direction. The destination file system is read/write and the source file
system is read-only, as shown in Figure 5 on page 16.

Source site Destination site


Celerra Network Server 1 Celerra Network Server 2

Data Mover Data Mover

Source Destination
file file
system system

Source Destination
SavVol SavVol

Source storage unit Destination storage unit

CNS-000761

Figure 5 Resynchronization

16 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Replication reversal
In this example, the reversal is used after a failover and resynchronization to
change the direction of the replication. The source site again accepts the source file
system updates from the network clients, and the replication service transfers these
updates to the destination site for playback to the destination file system, as shown
in Figure 6 on page 17.

Source site Destination site


Celerra Network Server 1 Celerra Network Server 2

Data Data
Mover Mover

Source Destination
file file
system system

Source Destination
SavVol SavVol

Source storage unit Destination storage unit

CNS-000766

Figure 6 Replication reversal

Note: A reversal requires both sites to be available and results in no data loss. During the
reversal phase, the source and destination file systems are set as read-only while the last
updates are transferred.

"Recovering replication data" on page 61 further describes the replication reversal


feature.

Communication between Celerra Network Servers


At the source and destination sites, you must build a trust relationship enabling
HTTP communication between Celerra Network Servers. This trusted relationship
is built upon a common passphrase set for both Celerra Network Servers. The 6- to
15-character passphrase is stored in clear text. It is used to generate a ticket for
Celerra-to-Celerra communication.

Using Celerra Replicator (V1) Version 5.6.47 17 of 172


Note: You use the nas_cel -create command to establish the relationship between your
Celerra Network Servers. For this command, ensure that the passphrase has at least 6
characters and no more than 15 characters. Remember that you must run the command on
both Celerra Network Servers and use the same passphrase.

The time on the Data Movers involved in a replication relationship and the Control
Stations at both sites must be synchronized with a maximum allowable skew of 10
minutes. Take into account time zones and daylight savings time, if applicable,
when using the Network Time Protocol (NTP) to synchronize the time.
Configuring EMC Celerra Time Services offers more information.
To establish communication, first, you must have root privileges and each site must
be active and configured for external communication. Table 1 on page 18 shows
information about the source and destination sites used in these examples.
Table 1 Source and destination sites information

Site Celerra name IP address

[source_site] cs100 192.168.168.114

[destination_site] cs110 192.168.168.102

Second, there must be IP network connectivity between both Control Stations.


Verify whether a relationship exists using the nas_cel -list command. If
communication is established, go to Task 3: "Create SnapSure checkpoint of
source file system" on page 42.

Note: This task is performed only for remote replication.

How resynchronization works


After a failover completes and the source site becomes operational, you can
resume replication using the -resync option. When the file systems are
resynchronizing, changes that occurred after the failover are copied to the source
site and replication is started but the replication direction is reversed. The
replication service is running on the destination site, and the playback service is
running on the source site. If this resynchronization is successful, you need not
perform a full file system copy.
Some reasons a resynchronization may not be possible are:
◆ You performed a failover but the system was unable to make the source file
system read-only. When the source site became available, it continued to
receive I/O to your source file system. If your replication service becomes
inactive, you must abort replication. You should not continue to allow I/O to the
original source file system.
◆ After you performed a failover, you decided to abort replication when the source
site became available.

18 of 172 Version 5.6.47 Using Celerra Replicator (V1)


If the replication service is active and you receive one of the following error
messages when attempting a resynchronization, you can run the -resync command
again by specifying the autofullcopy=yes option:
Error 2242: <file_system_name> : replication is not active, incremental resync not
possible. Abort and start replication or resync with autofullcopy=yes.
or
Resync copy failed. Incremental resync not possible. Abort and start replication or
resync with autofullcopy=yes.

Note: If this incremental resynchronization fails, restarting replication using a full file system
copy might take considerable time and resources. Plan carefully before using this option.

! !
CAUTION
Any data on the source file system not played back to the destination file system
prior to the failover is permanently lost.

How suspend works


Suspend is an option that allows you to temporarily stop an active replication
relationship and leave the replication in a condition that allows it to be restarted.
Suspend, when used in conjunction with the restart option, allows you to
temporarily stop replication, perform some action, and then restart the replication
relationship using an incremental rather than a full data copy.
After suspending a replication relationship, you can:
◆ Change the replication SavVol size. During the course of using replication, the
size of a SavVol may need changing because:
• The SavVol is too large and you want to reclaim the unused disk space.
• The SavVol is too small, which activates flow control.
◆ Mount the replication source or destination file system on a different Data
Mover.
◆ Change the IP addresses or interfaces the replication is using.
When you are restarting a replication relationship, you can specify a source
interface or allow the system to select it:
• If you specify an interface for the source site, replication uses that interface
until the user changes it.
• If you allow the system to select the interface, the interface can change to
keep the replication relationship running.
For example, if the network interface currently in use becomes unavailable, the
system attempts to select another interface. If it finds one, the replication
relationship continues to function.
The destination interface, regardless of how it is selected, is unchanged by the
system.

Using Celerra Replicator (V1) Version 5.6.47 19 of 172


When suspending a replication relationship, the system:
◆ Ensures all delta sets are transferred to the destination site.
◆ Plays back all outstanding delta sets.
◆ Creates a checkpoint on the source site, which is used to restart replication.

Note: The suspend checkpoint is named root_suspend_ckpt_xxx_1, where xxx is the ID of


the suspended replication relationship destination and 1 represents Celerra ID of a remote
replication relationship. No number appears for a local replication relationship.

The replication and playback services are no longer running after the suspend
action is complete.

How replication relationship restarts


The restart option allows you to restart a replication relationship by using an
incremental rather than a full data copy. Use the restart option when a replication
relationship is:
◆ Suspended
◆ Out-of-sync
Restart a suspended replication relationship
After you suspend a replication relationship using the -suspend option, only the -
restart option can restart it. This command verifies that the replication is in a
condition to allow a restart. It begins the process with an incremental copy using a
checkpoint of the source file system created when the replication suspended. You
must include this checkpoint when determining the maximum number of
checkpoints per file system used with replication.

Note: Before you restart a replication, make sure that all checkpoints are mounted.
Otherwise, a full data copy will be initiated instead of an incremental copy.

If you are using this procedure to:


◆ Increase the size of the replication SavVol, ensure that you specify the new
SavVol size using the savsize=<newsize> option.
◆ Change interfaces or IP addresses, specify them when you restart the
replication relationship.
When replication is restarted, default values are used for the replication policies.
For example, high water mark and timeout are set to 600. Specify new policies
when you restart replication using -option <options>.
Out-of-sync replication relationship
The source and destination file systems can become out-of-sync because:
◆ Network connectivity is lost.
◆ Incoming write rate is greater than the delta-set replay rate on the destination
file system.

20 of 172 Version 5.6.47 Using Celerra Replicator (V1)


This causes the following:
1. Delta sets accumulate on the source site until the SavVol is filled.
2. Changes are logged in memory until they can no longer be tracked.
3. The file systems fall out-of-sync if no source policy is specified.
Prevent an out-of-synchronization condition by specifying a source policy for a
replication relationship. Either stop accepting writes or stop all client access to the
file system.
Reestablishing the replication relationship
When the network connection is reestablished or the source file system is available
again, you can restart replication. To restart, you must have usable checkpoints of
the source file system. When restarting, all previous configuration parameters are
maintained and cannot be changed during the restart.
When restarting an out-of-sync replication relationship, the system:
◆ Determines the optimal checkpoint that meets the criterion of having a delta
number less than the delta-set number of the destination file system.
◆ Aborts the replication relationship.
◆ Restarts replication using the original configuration information.
◆ Performs an incremental copy of the file system using the appropriate
checkpoint.
If no valid checkpoint is available, you must abort and reestablish the replication
relationship. This means a complete (not an incremental) copy of the file system
must be done.
Checkpoint availability
To ensure that a valid checkpoint is available to restart an out-of-sync replication
relationship, create two checkpoints of the source file system. These restartable
checkpoints, named <source_fs_name>_repl_restart_1 and
<source_fs_name>_repl_restart_2, are automatically refreshed by a system CRON
job that runs every hour at 25 minutes after the hour. When a replication
relationship is out-of-sync, these checkpoints are available for the restart process.
You must include these two checkpoints when determining the maximum number of
checkpoints per file system used with replication.
Verify that these checkpoints are being refreshed by occasionally checking your
server log (server_log) file. If the system is not refreshing your checkpoints
regularly, call your service provider. The following is a sample message for a
successful refresh, where <fs_name> is the file system name:
Dec 29 07:29:23 2004 NASDB:7:13 refresh scheduled replication restart ckpt
<fs_name>_repl_restart_1 succeeded.

Using Celerra Replicator (V1) Version 5.6.47 21 of 172


System requirements for Celerra Replicator
This section details Celerra Network Server software, hardware, network, and
storage settings to use Celerra Replicator as described in this technical module.

Local replication
Table 2 on page 22 lists the system requirements for local replication.
Table 2 Local replication system requirements

• Celerra Network Server version 5.6.


Software
• Licenses for Celerra Replicator and SnapSure.

• One Celerra-storage (EMC Symmetrix® or EMC CLARiiON®) pair.


• CNS-14 or CFS-14 requires a minimum of a 510 Data Mover or later.
Hardware • To provide a graceful shutdown in an electrical power loss, Celerra Network
Server and the storage array need to have Uninterruptible Power Supply
(UPS) protection. If this is not provided, replication becomes inactive.

Network • IP addresses configured for the primary and secondary Data Movers.

• Sufficient storage space available for the source and destination file systems.
Storage • Sufficient SavVol space available for use by Celerra Replicator and
SnapSure.

Remote replication
Table 3 on page 22 lists the system requirements for remote replication.
Table 3 Remote replication system requirements (page 1 of 2)

• Celerra Network Server version 5.6 with the same Celerra version on the
source and destination Celerra Network Servers.

Note: For limited management capability and no fs_copy support: Celerra


Software Network Server version 5.5.39.2 on the source and 5.6.47 on the destination
Celerra Network Server.

• Licenses for Celerra Replicator and SnapSure.

• Minimum of two Celerra-storage (Symmetrix or CLARiiON) pairs.


• CNS-14 and CFS-14 require a minimum of one 510 Data Mover or later for
each Celerra-storage pair.
Hardware
• To provide a graceful shutdown in an electrical power loss, Celerra Network
Server and the storage array need to have UPS protection. If this is not
provided, replication becomes inactive.

22 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Table 3 Remote replication system requirements (page 2 of 2)

• IP addresses configured for the source and destination Data Movers (ports
8888 and 8887 used by replication for transferring data and internal
operations—contact Customer Service to change this setting).
• HTTPS connection between the Control Station on the source site and the
Network Control Station on the destination site (port 443—cannot be changed).
• Internet Control Message Protocol (ICMP) ensures that a destination
Celerra Network Server is accessible from a source Celerra Network Server.
The ICMP protocol reports errors and provides control data about IP packet
processing.

• Sufficient storage space available for the source and destination file systems.
Storage • Sufficient SavVol space available for use by Celerra Replicator and
SnapSure.

Consult "Configuration considerations" on page 35 for more information.

Using Celerra Replicator (V1) Version 5.6.47 23 of 172


Upgrading from previous Celerra Network Server
versions
This section describes the upgrade process when upgrading from a previous
version of Celerra Network Server to version 5.6 when using Celerra Replicator
(V1):
◆ Upgrade from a version earlier than 5.5.39.2
◆ Upgrade from version 5.5.39.2 or later (with out-of-family support)

Upgrade from a version earlier than 5.5.39.2


When upgrading from a previous version of Celerra Network Server that is earlier
than 5.5.39.2, to 5.6, note the following:
◆ Local replication
If you are upgrading a local replication, stop the playback service for all the
destination file systems by setting the timeout and high water mark to 0.
Upgrade Celerra Network Server, reset the timeout, and high water marks for
the destination file systems to their previous settings.
◆ Remote replication:
• If you are upgrading where only destination file systems are on the remote
system, upgrade Celerra Network Server at the destination site first, and
then upgrade the source site.
• You cannot upgrade a destination Celerra that is the target of a Celerra
running a 5.5 version earlier than 5.5.39.2.
• You cannot upgrade a Celerra running a version earlier than 5.5.39.2 that is
the source of an active replication session if the destination is also running a
5.5 version.
• You cannot upgrade a Celerra to 5.6 if there are bi-directional sessions
targeting a 5.5 Celerra as a destination.
◆ General replication considerations:
• Do not perform replication operations while upgrading (for example,
-failover, -resync, -reverse, -restart, -refresh).
• When upgrading the source and destination Celerra Network Servers from
version 5.5 to 5.6, replication sessions cannot be administered until both
sides have been upgraded to 5.6.

24 of 172 Version 5.6.47 Using Celerra Replicator (V1)


• Before starting an upgrade, ensure that no failed over or suspended
replication relationships are present. If you upgrade with replication
relationships in this state, these relationships will be unusable when the
upgrade completes.
• Introduced in version 5.4, SnapSure checkpoints use a new pageable
blockmap structure. Checkpoints used for replication are in this format.
Using SnapSure on EMC Celerra describes the pageable blockmap
structure in depth.
• Replication policies for time-out interval and high water mark are as follows:
– Time out: Acceptable timeout values are 0 or greater than 60 seconds up
to a limit of 24 hours.
– High water mark: The high water mark maximum value is 8000 MB. The
value should not exceed the size of the SavVol.

Note: You do not have to abort replication when upgrading Celerra Network Server
version.

Upgrade from Celerra Network Server version 5.5.39.2 or


later
Introduced in version 5.5.39.2, out-of-family replication support is recommended for
customers in a multi Celerra, edge-to-core environment who want to migrate to
NAS version 5.6 but want to upgrade their Celerras at different times without
interrupting data transfer.
In an edge-to-core configuration, where there are multiple source Celerras
replicating to one destination Celerra, you upgrade the destination (core) Celerra to
NAS version 5.6 first and then upgrade each of the source (edge) Celerras when
appropriate.
Out-of-family replication is not intended for use over an extended period of time as
there is limited replication management capability while in this environment. For
example, you cannot start a new replication, restart a suspended or inactive
replication, resync a failed-over replication, suspend or reverse a replication, or
perform a copy (full or differential) of a file system. Table 4 on page 26 provides the
support matrix for replication commands when in an out-of-family configuration. All
management commands are supported after both sides of the replication session
have been upgraded to the same NAS code family version.
When upgrading from version 5.5.39.2 or later of Celerra Network Server to 5.6,
note the following:
◆ Out-of-family replication support requires NAS version 5.5.39.2 or later on the
source Celerra and 5.6.47 or later on the destination Celerra.
◆ You cannot upgrade a 5.5.39.2 Celerra that is the source of an active replication
session if the destination is also running version 5.5.39.2.
◆ You cannot upgrade a Celerra to 5.6 if there are bi-directional sessions
targeting a 5.5 Celerra as a destination.

Using Celerra Replicator (V1) Version 5.6.47 25 of 172


◆ Out-of-family replication is unidirectional from a 5.5.39.2 source to a destination
running 5.6.47 or later. Replication from source 5.6 to destination 5.5 is not
supported.
◆ Does not support fs_copy.
◆ After upgrade, there is limited replication management support for Celerra
Replicator (V1) sessions running with different NAS code family versions
(source running 5.5.39.2 and destination running 5.6). Table 4 on page 26
provides the support matrix of replication management commands when in an
out-of-family configuration.
Table 4 Out-of-family replication command support matrix (page 1 of 2)

Replication state
Command
Active Inactive Suspended Failed-over Not configured

Abort on source Allowed Allowed NA Allowed NA

Abort on destination Allowed Allowed NA NA NA

Abort on both Not allowed Not allowed NA Not allowed NA

Refresh on source Allowed NA NA Allowed NA

Refresh on Allowed Allowed NA NA NA


destination

Refresh on both Not allowed NA NA Not allowed NA

Modify on source Allowed Allowed NA Allowed NA

Modify on destination Allowed Allowed NA NA NA

Modify on both Not allowed Not allowed NA Not allowed NA

Failover on Allowed Allowed NA NA NA


destination (default | (default | now) (default | now)
Now | Sync) Not Allowed Not Allowed
(sync) (sync)

Resync on NA NA NA Not allowed NA


destination

Reverse on source Not allowed Not allowed NA NA NA

Restart on source NA Not allowed Not allowed NA NA

Suspend on source Not allowed Not allowed NA NA NA

List from source Allowed Allowed NA Allowed NA

List from destination Allowed Allowed NA NA NA

26 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Table 4 Out-of-family replication command support matrix (page 2 of 2)

Replication state
Command
Active Inactive Suspended Failed-over Not configured

Info from source Allowed Allowed NA Allowed NA

Info from destination Allowed Allowed NA NA NA

Start from source NA Not allowed NA NA Not allowed

fs_copy from source NA NA NA NA Not allowed

Prerequisites
Before upgrading, make sure that the Celerra to be updated:
◆ Is running version 5.5.39.2 or later.
◆ If Celerra to be upgraded is a destination, make sure that the source hosting the
active replication sessions is running version 5.5.39.2 or later.
◆ If Celerra to be upgraded is a source, make sure that the destination hosting the
active replication sessions is running version 5.6.47 or later.
◆ Has no fs_copy sessions running.
◆ Is running under minimum load. If there is a high rate of I/O during upgrade it
may cause replications to become inactive.
◆ Is not hosting both the source and destination sides of active replications
running version 5.5. If bi-directional sessions exist on the Celerra to be
upgraded, do the following:
a. Suspend all replication sessions in one direction. (Either the sessions
running from A to B or the sessions running from B to A.)
b. Upgrade the Celerra that is hosting only the destination side of the active
replication sessions.
c. Data transfer will continue but with limited replication management capability.
d. Upgrade the source Celerra.
e. Restart all the suspended replication sessions.
Procedure
To upgrade to Celerra Network Server version 5.6.47:

Step Action

1. Upgrade the destination Celerra from NAS code version 5.5 to 5.5.39.2.

2. Upgrade all of the source Celerras from NAS code version 5.5 to 5.5.39.2.

Using Celerra Replicator (V1) Version 5.6.47 27 of 172


Step Action

3. Upgrade the destination Celerra from NAS code version 5.5.39.2 to 5.6.47.
Data continues to transfer between the source and destination sites, but there is limited
replication management capability until you upgrade the source Celerra to 5.6.47. For
example, you cannot start a new replication, restart a suspended or inactive replication,
resync a failed-over replication, suspend or reverse a replication, or perform a copy (full or
differential) of a file system. Table 4 on page 26 provides the out-of-family replication
command support matrix.

4. Upgrade the source Celerra from 5.5.39.2 to 5.6.47.

5. Repeat step 4 for all of the source Celerras to be upgraded.

28 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Planning considerations for Celerra Replicator
Before you use Celerra Replicator:
◆ Review replication policies
◆ Determine SavVol size
◆ Determine the number of replications per Data Mover
◆ Review the configuration considerations

Replication policies
Most replication policies can be established for one replication relationship (using
the fs_replicate command) or all replication relationships on a Data Mover by
setting a parameter. Celerra Replicator has policies to:
◆ Control delta-set generation using a time-out interval and high water mark.
◆ Control how to handle data if network connectivity is lost (flow control). "Celerra
Replicator flow control" on page 31 describes this action.
◆ Set the maximum IP bandwidth size used by a replication session. "Set
bandwidth size" on page 109 details this policy.
◆ Set the data amount to be sent across the IP network before an
acknowledgment is required from the receiving side. This is controlled by a
parameter for the TCP window size (tcpwindow). "Accommodating network
concerns" on page 127.
IP Alias with IP replication
Celerra Network Server versions 5.5.28.1 and later support IP Alias with IP
replication. All restrictions on Control Station failover also apply to IP Alias with IP
replication configurations. The following guidelines apply to this feature:
◆ When using IP replication for the first time, or on new systems, configure IP
Alias first, and use IP Alias in the -ip <ipaddr> option of nas_cel -create
command.
◆ For existing systems with existing IP replication sessions, the current slot_0 IP
address (primary Control Station IP address) must be used. For example:
nas_config -IPalias -create 0 -> Do you want slot_0 IP address as your alias
[yes or no] yes
◆ If the Control Station fails over while IP replication is running, the IP replication
command (fs_replicate) might need to be re-executed manually. Check logs
(/nas/log/cmd_log*, server_log command output etc.) to determine how to
proceed. Keep the fs_replicate command output for resync, suspend, restart,
failover, and reverse options in case of failure, then execute the rest of steps
based on the instruction in command output.
◆ When IP Alias is deleted using the nas_config -IPalias -delete command, the IP
address of the primary or the secondary Control Station is not changed.
Changes to the IP address of the primary or the secondary Control Station must
be done separately. IP replication depends on communication between the

Using Celerra Replicator (V1) Version 5.6.47 29 of 172


source Control Station and the destination Control Station. When IP Alias is
used for IP replication, deleting the IP Alias breaks the communication. The IP
address which was used as the IP Alias must be restored on the primary
Control Station to restore the communication.
◆ While performing a Celerra code upgrade using ssh or telnet, do not use an IP
Alias in the ssh or telnet to log in to the Control Station.
Control delta-set generation
A delta set contains the block modifications made to the source file system and is
used by the replication service to periodically synchronize the destination file
system with the source file system. The amount of information within the delta set is
based on the source file system’s activity and how you set the replication policies,
time out, and high water mark values. The minimum delta-set size is 128 MB.
The replication service is triggered by either the time-out policy or the high water
mark policy, whichever is reached first. However, when the maximum delta-set size
is reached (8 GB), a new delta set is generated, regardless of the time-out policy:
◆ Time-out policy (where the system either generates or plays back a delta set):
• Source site: The interval at which the replication service automatically
generates a delta set (for example, every 1200 seconds).
• Destination site: The interval at which the playback service automatically
plays back all available delta sets to the destination file system (for example,
every 600 seconds).
At both sites, the default time-out value is 600 seconds. Acceptable time-out
values are 0 or greater than 60 seconds up to a limit of 24 hours. A value of 0,
indicating there is never a timeout, pauses the replication activities for this
policy.
◆ High water mark:
• Source site: The size in MB of the file system changes accumulated since
the last delta set, at which the replication service automatically creates a
delta set on the SavVol. For example, when the amount of change reaches
1200 MB in size, a delta set is generated.
• Destination site: The size in MB of the delta sets present on the destination
SavVol, at which the replication service automatically plays back all available
delta sets to the destination file system. For example, when the amount of
change reaches 600 MB in size, playback occurs.
At both sites, the default high water mark value is 600 MB. Acceptable high
water mark values are 0 or greater than 30 MB up to a maximum value of 8 GB.
This value should not exceed the size of the SavVol. A value of 0 pauses the
replication activities and disables this policy.

Note: A delta set may not be generated or copied if a flow control is triggered. "Celerra
Replicator flow control" on page 31 provides further information.

30 of 172 Version 5.6.47 Using Celerra Replicator (V1)


How time-out and high water mark policies work
How time-out and high water mark policies work discusses how these replication
policies work for the source and destination file systems.
Replication policies for the source file system and when one of these replication
policies is triggered, the following occurs:
1. The replication service automatically generates a delta set from the
accumulated changes to the file system and stores it in the SavVol. Each delta
set is recorded and processed per replication policy trigger. Each delta set
contains a set of changes to the source file system that occur since the creation
of the previous delta set.
2. Replication tracks all subsequent changes made to the source file system.
3. The delta sets are transferred from the SavVol on the source site to the SavVol
on the destination site (for remote replication), and updates the destination file
system through the replication playback service (for local and remote
replication).
4. The replication service waits for the next event to create the next delta set.
Replication policies for the destination file system:
◆ The playback service continually polls the SavVol on the destination site to play
back each delta set to the destination file system, synchronizing it with the
source file system. This playback rate is based on the specified replication
policy. After the delta set is copied to the destination file system, the next delta
set is processed.
◆ For optimization, delta sets available before a trigger is reached are merged into
one active playback session.
Celerra Replicator flow control
Celerra Replicator flow control activates when the playback rate on the destination
site is too slow, when network connectivity is lost, or when the write rate from the
source site is too fast.
Celerra Replicator activates flow control in an attempt to allow the network to catch
up. In most cases, Celerra Replicator should not be in a flow control state. Flow
control is activated:
◆ If a delta set that has not been replayed on the destination file system is about
to be overwritten by a newer delta set, the destination file system temporarily
holds the data flow until the delta set is replayed. This happens when the delta-
set playback rate for the destination file system is too slow to handle the source
file system updates.
◆ When a delta set cannot be transferred from the SavVol on the source site to
the SavVol on the destination site (for example, the network is unavailable), the
replication service stops the data flow to the destination site. During this time,
Celerra Replicator tracks the source file system modifications, and continually
retries connectivity to the destination site. If the network is down, a message is
sent to the system log on the source site.
◆ When the SavVol at the source site is full, the replication service suspends
copying the modified blocks to the SavVol. This prevents the overwriting of delta

Using Celerra Replicator (V1) Version 5.6.47 31 of 172


sets not yet transferred to the SavVol on the destination site. During this time,
modifications to the source file system are still tracked, but are not copied to the
SavVol until it has available space.
Celerra Network Server tracks changes in memory only until these changes
represent a delta set the same size as the Celerra Replicator SavVol. If the system
can no longer track changes in memory, the system behaves in one of three ways,
as explained in Table 5 on page 32.
Table 5 Replication flow-control options

How to specify for all


How to specify for a single
Behavior replication sessions on a
replication session
Data Mover

Temporarily halts all I/Os to the Set an option using Set the VRPL freeze
source file system until fs_replicate -modify parameter as described in the
sufficient space is available on -option autofreeze=yes. "How EMC Celerra Network Server
the source SavVol. During this time-out and high water mark Parameters Guide.
time, the file system is policies work" on page 31
inaccessible to network clients. describes this policy.
When space is available on
the Celerra Replicator SavVol,
the source file system is
mounted and begins accepting
I/Os.

Temporarily stops writing data Set an option using Set the VRPL read-only
to the source file system by fs_replicate -modify parameter as described in the
mounting it read-only. Users -option autoro=yes. "How time- EMC Celerra Network Server
still have read-only access to out and high water mark Parameters Guide.
the source file system. policies work" on page 31
When space becomes describes this policy.
available on the Celerra
Replicator SavVol, the source
file system is remounted
read/write and begins
accepting all I/Os to the source
file system.

Allows the replication service "Restarting a replication Not applicable.


to fall out of sync. relationship" on page 89
describes how to restart
replication after it has fallen
out of sync.

As shown in Table 5 on page 32, these policies can be set for one replication
session or for all replication sessions on a Data Mover. If there is a conflict between
these two policies, the one defined for a single replication session takes
precedence.
You can set up alerts to notify you when these events occur. "Events for Celerra
Replicator" on page 121 details Celerra Replicator events, and Configuring EMC
Celerra Events and Notifications describes how to use them.

32 of 172 Version 5.6.47 Using Celerra Replicator (V1)


SavVol size requirements for remote replication
Replication SavVols store changes that occurred on the source file system not yet
replayed to the destination file system. Celerra Replicator requires a SavVol at
source and destination sites. The minimum SavVol size is 1 GB and the maximum
is 500 GB. Determine the size of these SavVols based on the following criteria:
◆ Investigate the source file system size, update ratio of the source file system
per day, and the WAN network bandwidth for the source/destination file systems
between the source and destination Celerra Network Servers. Use nas_fs
-size to calculate the SavVol size and nas_disk -list to find the entire file system
size. The EMC Celerra Network Server Command Reference Manual provides
more command details.
◆ Evaluate the risk tolerance to network outages. For example, if the network
experiences long outages, such as two days, ensure that the SavVol on the
source site will allow capturing two days of delta sets.
◆ If performing replication on multiple file systems per Data Mover, consider the
available network bandwidth per file system.
◆ Determine whether the network bandwidth is sufficient to transfer the changed
data from the source to destination file system. If the rate of change on the
source is continuously greater than available network bandwidth, the replication
service will not transfer data quickly enough, eventually becoming inactive.
You may also contact EMC Customer Service or read the E-Lab Interoperability
Navigator on Powerlink to size the SavVols.
Change replication SavVol requirements
In most cases, the default size of the SavVol is sufficient (10 percent of the source
file system size). To accommodate a large network outage or allow for brief periods
where the incoming change rate significantly exceeds the ability of the network to
send changes to the destination site or both, consider increasing your replication
SavVol size.
For example, a 50 GB SavVol is sufficient for a 500 GB file system that incurs 20
GB of change per day to cover approximately two and one-half days of network
outage without any flow-control events. But to cover a longer outage period, you
can enlarge the SavVol.

Using Celerra Replicator (V1) Version 5.6.47 33 of 172


Table 6 on page 34 describes changing the replication SavVol size and explains
when to use each method.
Table 6 Changing Celerra Replicator SavVol size

When to use What to do Procedure

Before starting any Change default size of each "Change the Celerra
replication processing. Celerra Replicator SavVol Replicator SavVol default
from 10% of the source file size" on page 123
system.
By default, the system
allocates 10% of the size of
the source file system for the
replication SavVol on the
source and destination sites.

At the start of a replication Control SavVol size for a file Task 6: "Begin replication" on
instance. system by specifying a page 48
specific SavVol size. Use
savsize option for
fs_replicate -start.

After replication is running. Revise SavVol size to meet "Suspend a replication


your changing requirements. relationship" on page 81

Determine the number of replications per Data Mover


Determine the maximum number of replication sessions per Data Mover based on
your configuration (such as the WAN network bandwidth, delta-set update ratio, and
the production I/O workload). This number is also affected by whether you are
running SnapSure and Celerra Replicator on the same Data Mover. Both of these
applications share the available memory on a Data Mover.
For all configurations, there is an upper limit to the number of replications allowed
per Data Mover. E-Lab Interoperability Navigator on Powerlink details the current
number of replications allowed per Data Mover.

Note: If you plan to run loopback replications, remember that each loopback replication
counts as two replication sessions because each session encapsulates outgoing and
incoming replications.

To learn the number of replication sessions per Data Mover:


1. Determine the SavVol size of each replication, as described in "SavVol size
requirements for remote replication" on page 33.
2. Verify that the total storage on a Data Mover (including any source and
destination file systems and associated SavVols) does not exceed the
guidelines for that Data Mover. These guidelines are detailed in the E-Lab
Interoperability Navigator on Powerlink.
3. Celerra Replicator should be transferring delta sets faster than it creates them.
Verify that the delta sets per Data Mover can be transferred to the destination

34 of 172 Version 5.6.47 Using Celerra Replicator (V1)


site with the available WAN network bandwidth. An active flow-control condition
may be indicative of this situation.

Note: To provide a stable network transfer rate for delta sets, it is strongly
recommended that you configure a dedicated network port for Data Mover transfers.

4. Verify that the Data Mover can handle all replication sessions and production
I/Os. You can also monitor memory usage and CPU usage using the
server_sysstat command. This command shows total memory utilization, not
just Celerra Replicator and SnapSure memory usage.

Note: Use Celerra Manager to monitor memory and CPU usage by creating a new
notification on Celerras > Notifications > Data Mover Load tab.

Contact EMC Customer Service for additional advice.

Configuration considerations
Before setting the replication policy triggers, consider the following:
◆ To avoid the source and destination file systems from becoming out-of-sync, do
not allow the replication service to create delta sets significantly faster than it
can copy them to the destination file system. Set the delta-set creation
replication policy to a higher number (for example, 1200 seconds) than the
delta-set playback number (for example, 600 seconds). The replication policies
you establish for creating and replaying delta sets depend on the size and
number of transactions processed on the source file system.
◆ Determine if the network bandwidth can effectively transport the production
change data generated at the source site to the destination site.
◆ During the delta-set playback on the destination file system, network clients can
access the destination file system. However, at the beginning of the delta-set
playback for CIFS (Common Internet File Service) clients, there is a temporary
freeze/thaw period that may cause a network disconnect. As a result, do not set
the replication policy to a low number because this reduces the availability of
the destination file system.
To eliminate this freeze/thaw period, create a checkpoint of the destination file
system and mount it for client access at the destination site. However, this
checkpoint will not contain the most up-to-date production data.
◆ Carefully evaluate the infrastructure of the destination site by reviewing items
such as:
• Subnet addresses
• Unicode configuration
• Availability of name resolution services; for example, WINS, DNS, and NIS
• Availability of WINS/PDC/BDC/DC in the correct Microsoft Windows NT, or
Windows Server domain
• Share names

Using Celerra Replicator (V1) Version 5.6.47 35 of 172


• Availability of user mapping (for example, using EMC Usermapper for
Celerra systems). The CIFS environment requires more preparation to set
up a remote configuration because of the higher demands on its
infrastructure than the network file system (NFS) environment (for example,
authentication is handled by the domain controller). For the CIFS
environment, you must perform mappings between the usernames/groups
and UIDs/GIDs with EMC usermapper or local group/password files on the
Data Movers.

Note: Replicating EMC Celerra CIFS Environments (V1) describes configuration


considerations in depth.

◆ Local groups are not supported on replicated file systems unless you use
VDMs. Replicating EMC Celerra CIFS Environments (V1) describes this
consideration more fully.
◆ The replication SavVol for the delta sets must be large enough to store and
process all the delta-set write I/Os, and the SnapSure SavVol for the
checkpoints must be able to store all the source file system block changes for
the initial synchronization.
◆ The destination file system can only be mounted on one Data Mover, even
though it is read-only.
◆ At the application level, as well as the operating system level, some
applications might have limitations on the read-only destination file system due
to caching and locking.
◆ If you are planning on enabling international character sets (Unicode) on your
source and destination sites, you must first set up translation files on both sites
before starting Unicode conversion on the source site. Using International
Character Sets with EMC Celerra covers this consideration.
◆ Celerra FileMover feature supports replicated file systems. This is described in
Using EMC Celerra FileMover.
◆ Celerra File-Level Retention Capability supports replicated file systems. Using
File-Level Retention on EMC Celerra provides additional configuration
information.

36 of 172 Version 5.6.47 Using Celerra Replicator (V1)


User interface choices for Celerra Replicator
Celerra Network Server offers flexibility in managing networked storage based on
your support environment and interface preferences. This technical module
describes how to set up and manage replication using the command line interface
(CLI). You can also perform most tasks using Celerra Manager — Basic Edition.
The following documents provide additional information about managing Celerra:
◆ Getting Started with Celerra details user interface choices.
◆ Learning about EMC Celerra on the EMC Celerra Network Server
Documentation CD and the application’s online help system describe each
application’s capabilities.
◆ The EMC Celerra Network Server Release Notes provide additional, late-
breaking information about Celerra management applications.

Using Celerra Replicator (V1) Version 5.6.47 37 of 172


Roadmap for Celerra Replicator
This section lists the tasks for configuring and managing Celerra Replicator.
Celerra Replicator configuration tasks:
1. "Initiating replication" on page 39
2. "Recovering replication data" on page 61
Celerra Replicator management tasks:
◆ "Abort Celerra Replicator" on page 79
◆ "Suspend a replication relationship" on page 81
◆ "Restarting a replication relationship" on page 89
◆ "Extending the size of a file system" on page 98
◆ "Resetting replication policy" on page 105
◆ "Reverse the direction of a replication relationship" on page 111
◆ "Monitor replication" on page 114
◆ "Checking playback service and outstanding delta sets" on page 115
◆ "Events for Celerra Replicator" on page 121
◆ "Change the Celerra Replicator SavVol default size" on page 123
◆ "Change the passphrase between Celerra Network Servers" on page 124
◆ "Managing and avoiding IP replication problems" on page 125
◆ "Transporting replication data using disk or tape" on page 139
◆ "Setting up the CLARiiON disk array" on page 147

38 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Initiating replication
In most cases, you will have a functioning NFS or CIFS environment before you use
Celerra Replicator. If you do not, ensure that you set up the following on the source
and destination Celerra Network Servers before using Celerra Replicator:
◆ Establish the IP infrastructure.
◆ Establish name service. Read Configuring EMC Celerra Naming Services for
more information about establishing name service.
◆ Synchronize the time between the Data Movers and Control Stations involved in
the replication relationship. The maximum allowable time skew is 10 minutes.
Read Configuring EMC Celerra Time Services for more information.
◆ Establish user mappings. Read Configuring EMC Celerra User Mapping for
more information on establishing user mappings.
The process of setting up a local or remote replication relationship assumes the
following:
◆ The source file system is created and mounted as read/write on a Data Mover.
◆ The destination file system is not created.
◆ The Celerra Network Server version of the destination site must be same as the
version of the source site.

Note: The communication between Celerra Control Stations uses HTTPS.

When using remote replication, it is useful to create SnapSure checkpoints of the


source file system to restart the source and destination sites if they fall out-of-sync.
After you create the checkpoints, the system automatically keeps them up to date,
as described in "Out-of-sync replication relationship" on page 20.
To set up replication you have to complete the following tasks:
1. "Establish communication" on page 40
2. "Verify communication" on page 40
3. "Create SnapSure checkpoint of source file system" on page 42
4. "Create the destination file system" on page 45
5. "Copy checkpoint to the destination file system" on page 45
6. "Begin replication" on page 48
7. "Create a second checkpoint of the source file system" on page 50
8. "Copy incremental changes" on page 52
9. "Verify file system conversion" on page 54
10. "Check replication status" on page 55
11. "Create restartable checkpoints" on page 59

Using Celerra Replicator (V1) Version 5.6.47 39 of 172


Note: The first two tasks are used for remote replication and do not apply to setting up a
local replication relationship. The commands for setting up a local replication relationship
are included for your reference. The output reflects a remote replication.

Task 1: Establish communication


To establish communication:

Action

To establish a trusted relationship at each site, logged in as root, use this command syntax:
# nas_cel -create [-name <cel_name>] <ip> -passphrase <passphrase>
where:
<cel_name> = name of the remote (destination) Celerra Network Server in the configuration
<ip> = IP address of the remote Control Station in slot 0
<passphrase> = secure passphrase used for the connection, which must have 6- to 15-
characters and be the same on both sides of the connection
Example:
To set up a trust relationship, type the following commands at both sites:
[source_site]# nas_cel -create eng25271 -ip 172.24.252.71 -passphrase
nasadmin
[destination_site]# nas_cel -create eng25246 -ip 172.24.252.46 -
passphrase nasadmin

Note: If you need to change the passphrase later, follow the procedure described in "Change the
passphrase between Celerra Network Servers" on page 124.

Output

From source site eng25271, to set up relationship with destination site eng25246:
operation in progress (not interruptible)...
id = 3
name = eng25271
owner = 0
device =
channel =
net_path = 172.24.252.71
celerra_id = 0001901003890010
passphrase = nasadmin

Task 2: Verify communication


This task is performed only for remote replication. To verify communication:
◆ "Verify communication at the source site" on page 41
◆ "Verify communication at the destination site" on page 41
◆ "View passphrase" on page 42

40 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Verify communication at the source site
At the source site, check whether Celerra Network Servers can communicate with
one another.

Action

To verify that the source and destination sites can communicate with each other, type the
command at each site:
[source_site]$ nas_cel -list

Output

id name owner mount_dev channel net_path CMU


0 cs100 0 192.168.168.114 APM000340000680000
1 eng168123 201 xxx.xxx.xxx.xxx APM000437048940000
3 eng16853 501 xxx.xxx.xxx.xxx 0001835017370000
5 cs110 503 192.168.168.102 APM000446038450000

Note

The sample output shows the source site can communicate with the destination site, cs110.

Verify communication at the destination site


At the destination site, check whether Celerra Network Servers can communicate
with one another.

Action

To verify that the source and destination sites can communicate with each other, at each site, type:
[destination_site]$ nas_cel -list

Output

id name owner mount_dev channel net_path CMU


0 cs110 0 192.168.168.102 APM000446038450000
2 cs100 501 192.168.168.114 APM000340000680000

Note

The sample output shows the destination site can communicate with the source site, cs100.

Using Celerra Replicator (V1) Version 5.6.47 41 of 172


View passphrase
The 6- to 15-character passphrase is used to authenticate with a remote Celerra
Network Server.

Action

To view the passphrase of a Celerra Network Server, use this command syntax:
$ nas_cel -info id=<cel_id>
where:
<cel_id> = Celerra ID

Note: Celerra ID is assigned automatically. To view this ID of a remote system, use the
nas_cel -list command. You can also use the hostname.

Example:
To view the passphrase of the Celerra system, type:
$ nas_cel -info id=5

Output

id = 5
name = cs110
owner = 503
device =
channel =
net_path = 192.168.168.102
celerra_id = APM000446038450000
passphrase = nasadmin

Task 3: Create SnapSure checkpoint of source file system


A SnapSure checkpoint is used as the baseline of data to copy to the destination
file system. Copying this baseline data from the source to the destination site over
the IP network can be a time-consuming process.
When using remote replication, you can use an alternate method to copy the initial
checkpoint of the source file system. You can back it up to a disk array or tape drive
and transport it to the destination site. To use this alternate method, go to
"Transporting replication data using disk or tape" on page 139 to continue with this
procedure. Using SnapSure on EMC Celerra provides details on SnapSure
checkpoints.

! !
CAUTION
◆ When creating checkpoints, be careful not to exceed your system’s limit. Celerra
permits 96 checkpoints per PFS, regardless of whether the PFS is replicated, for
all systems except the Model 510 Data Mover (which permits 32 checkpoints with
PFS replication and 64 checkpoints without). This limit counts existing
checkpoints, or those already created in a schedule and might count two
restartable checkpoints as well as a third checkpoint created by certain
replication operations on either the PFS or SFS.
◆ If you are at the limit, delete existing checkpoints to create space for newer
checkpoints, or do not create new checkpoints if existing ones are more

42 of 172 Version 5.6.47 Using Celerra Replicator (V1)


important. Be aware when you start to replicate a file system, the facility must
create two checkpoints, otherwise replication will not start. For example, if you
have 95 checkpoints and want to start a replication, the 96th checkpoint will be
created, but replication will fail when the system tries to create the 97th
checkpoint because the limit is breached.
◆ Also, when scheduling, be careful not to keep any checkpoints that will surpass
the limit otherwise you can not start a replication. In other words, if all
checkpoints you specify to keep are created, they must be within the limit.

Action

To create a SnapSure checkpoint of the source file system, use this command syntax:
$ fs_ckpt <fs_name> -Create
where:
<fs_name> = file system name on which a checkpoint is created
Remote replication example:
To create a checkpoint of the source file system src_ufs1, type:
$ fs_ckpt src_ufs1 -Create
Local replication example:
To create a checkpoint of the source file system local_src, type:
$ fs_ckpt local_src -Create

Note: The output shown is for the remote replication example.

Using Celerra Replicator (V1) Version 5.6.47 43 of 172


Output

operation in progress (not interruptible)...id = 88


name = src_ufs1
acl = 0
in_use = True
type = uxfs
worm = off
volume = v243
pool = clar_r5_performance
member_of = root_avm_fs_group_10
rw_servers= server_2
ro_servers=
rw_vdms =
ro_vdms =
ckpts = src_ufs1_ckpt1
stor_devs = APM00034000068-001F,APM00034000068-001E
disks = d21,d15
disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14
server=server_2
disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13
server=server_2

id = 90
name = src_ufs1_ckpt1
acl = 0
in_use = True
type = ckpt
worm = off
volume = vp246
pool = clar_r5_performance
member_of =
rw_servers=
ro_servers= server_2
rw_vdms =
ro_vdms =
checkpt_of= src_ufs1 Mon Feb 7 06:58:10 EST 2005
used = 1%
full(mark)= 90%
stor_devs = APM00034000068-001F,APM00034000068-001E
disks = d21,d15
disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14
server=server_2
disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13
server=server_2

44 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Task 4: Create the destination file system
The destination file system is created as rawfs because replication starts on a rawfs
file system. When all steps to set up replication are complete, the file system will be
a normal uxfs type. The destination file system must be the same size as the
source file system. Managing Celerra Volumes and File Systems Manually and the
appropriate man page describe how to create and use a file system.

Step Action

1. Create a destination file system using the samesize= option by typing:


$ nas_fs -name dst_ufs1 -type rawfs -create
samesize=src_ufs1:cel=cs100 pool=clar_r5_performance
Local replication example:
$ nas_fs -name local_dst -type rawfs -create samesize=local_source
pool=clar_r5_performance

2. Create a mount point on the destination Data Mover by typing:


$ server_mountpoint server_2 -create /dst ufs_1
Local replication example:
$ server_mountpoint server_2 -create /local_dst

3. Mount the file system as read-only on the destination Data Mover by typing:
$ server_mount server_2 -option ro dst_ufs1 /dst_ufs1
Local replication example:
$ server_mount server_2 -option ro local_dst /local_dst

Note: The destination file system can only be mounted on one Data Mover, even though it
is read-only.

Task 5: Copy checkpoint to the destination file system


This task copies the entire checkpoint of the source file system created in Task 3:
"Create SnapSure checkpoint of source file system" on page 42 to the destination
file system. This creates a baseline copy of the source file system on the
destination file system. This copy is updated incrementally with changes occurring
to the source file system.
Perform this copy task once per file system that is to be replicated. The checkpoint
must be copied without converting it to uxfs, by using the convert=no option.
Use of the monitor=off option runs this command as a background process,
allowing you to run several copy sessions simultaneously.

Note: If the primary file system extends during the running of the fs_copy command and
before replication starts in Task 6: "Begin replication" on page 48, you must extend the
destination file system manually to keep file system sizes identical. Use the nas_fs -xtend
command.

Using Celerra Replicator (V1) Version 5.6.47 45 of 172


Action

To copy a checkpoint to the destination file system, use this command syntax:
$ fs_copy -start <srcfs> <dstfs>:cel=<cel_name> -option
convert=no,monitor=off
where:
<srcfs> = source file system checkpoint.
<dstfs> = destination file system.
<cel_name> = destination Celerra Network Server name.
-option convert=[yes|no] = allows the conversion of the <dstfs> to uxfs after the file system
copy is executed. If no is specified, when the copy has completed, the <dstfs> remains a rawfs file
system type. The default is yes.
-option monitor=off = progress of the copy is printed to the screen by default. The off
option forces the command to run as a background process.
Remote replication example:
To copy the checkpoint, src_ufs1_ckpt1, to the destination file system without converting it to uxfs,
type:
$ fs_copy -start src_ufs1_ckpt1 dst_ufs1:cel=cs110 -option
convert=no,monitor=off
Local replication example:
To copy the checkpoint, local_src_ckpt1, to the destination file system, local_dst, type:
$ fs_copy -start local_src_ckpt1 local_dst -option convert=no,
monitor=off

Note: The output shown is for the remote replication example.

Output

operation in progress (not interruptible)...id = 90


name = src_ufs1_ckpt1
acl = 0
in_use = True
type = ckpt
worm = off
volume = vp246
pool = clar_r5_performance
member_of =
rw_servers=
ro_servers= server_2
rw_vdms =
ro_vdms =
checkpt_of= src_ufs1 Mon Feb 7 06:58:10 EST 2005
used = 1%
full(mark)= 90%
stor_devs = APM00034000068-001F,APM00034000068-001E
disks = d21,d15
disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14
server=server_2
disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13
server=server_2

46 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Output

id = 126
name = dst_ufs1
acl = 0
in_use = True
type = rawfs
worm = off
volume = v272
pool = clar_r5_performance
member_of = root_avm_fs_group_3:cs110
rw_servers=
ro_servers= server_2
rw_vdms =
ro_vdms =
backup_of = src_ufs1 Mon Feb 7 06:58:10 EST 2005
stor_devs = APM00044603845-0008,APM00044603845-0007
disks = d8,d9
disk=d8 stor_dev=APM00044603845-0008 addr=c0t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c32t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c16t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c48t1l2
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c16t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c48t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c0t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c32t1l1
server=server_2

done

Verify the copied checkpoint to the destination file system


Action

To verify that the fs_copy command completes, type:


$ fs_copy -list

Output

Local Source Filesystems


Id Source Destination Status %Remaining CommState

Local Destination Filesystems


Id Source Destination Status %Remaining CommState

Note

The fs_copy session is not listed in the output indicating the copy is complete.

Using Celerra Replicator (V1) Version 5.6.47 47 of 172


Task 6: Begin replication
When you start replication, the system verifies that primary and secondary Data
Movers can communicate with each other. Next, it starts replicating and then begins
tracking all changes made to the source file system. You start this process once per
file system to be replicated.
Set your replication policies when you establish this replication relationship.
"Replication policies" on page 29 describes this feature. If you want to specify a
specific interface or IP address for the replication relationship, do so when you start
replication. If you specify an interface for the source site, replication uses that
interface until it is changed by the user. If you allow the system to select the
interface, it can change to keep the replication relationship running. For example, if
the network interface currently being used becomes unavailable, the system
attempts to select another interface. If one is found, the replication relationship
continues to function. The destination interface, regardless of how selected, is
unchanged by the system.
Any future changes to this information requires suspending and restarting
replication, as detailed in "Suspend a replication relationship" on page 81 and
"Restarting a replication relationship" on page 89.

Action

To start replication for the first time, use this command syntax:
$ fs_replicate -start <srcfs> <dstfs>:cel=<cel_name>
where:
<srcfs> = source file system
<dstfs> = destination file system
<cel_name> = destination Celerra Network Server

Note: Multiple fs_replicate -start processes must be executed sequentially, not in parallel. Run
only one fs_replicate -start command at a time.

Remote replication example:


To start replication for source file system src_ufs1 and a destination file system dst_ufs1, type:
$ fs_replicate -start src_ufs1 dst_ufs1:cel=cs110
Local replication example:
To start replication for source file system local_src and a destination file system local_dst, type:
$ fs_replicate -start local_src local_dst

Note: The output shown is for the remote replication example.

48 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Output

operation in progress (not interruptible)...id = 88


name = src_ufs1
acl = 0
in_use = True
type = uxfs
worm = off
volume = v243
pool = clar_r5_performance
member_of = root_avm_fs_group_10
rw_servers= server_2
ro_servers=
rw_vdms =
ro_vdms =
ckpts = src_ufs1_ckpt1
ip_copies = dst_ufs1:cs110
stor_devs = APM00034000068-001F,APM00034000068-001E
disks = d21,d15
disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14
server=server_2
disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13
server=server_2

id = 126
name = dst_ufs1
acl = 0
in_use = True
type = rawfs
worm = off
volume = v272
pool = clar_r5_performance
member_of = root_avm_fs_group_3:cs110
rw_servers=
ro_servers= server_2
rw_vdms =
ro_vdms =
backup_of = src_ufs1 Mon Feb 7 06:58:10 EST 2005
stor_devs = APM00044603845-0008,APM00044603845-0007
disks = d8,d9
disk=d8 stor_dev=APM00044603845-0008 addr=c0t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c32t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c16t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c48t1l2
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c16t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c48t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c0t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c32t1l1
server=server_2
done

Using Celerra Replicator (V1) Version 5.6.47 49 of 172


Note

The system selects the interface if you do not specify one.


Error messages:
• If you receive an error message stating the interface is not configured or is invalid, the IP
addresses for the interface ports are not configured on the destination site. Define these
interface ports by running the server_ifconfig command at the destination site.
• If Error 2211: Sec: invalid id specified appears, the local and destination sites have different
passphrases. To modify this, follow the procedure in "Change the passphrase between Celerra
Network Servers" on page 124.

Task 7: Create a second checkpoint of the source file system


The changes between the two checkpoints are copied to the destination file system
in the next task.

Action

To create a second checkpoint of the source file system, which is compared to the initial
checkpoint, use this command syntax:
$ fs_ckpt <fs_name> -Create
where:
<fs_name> = file system name for which a checkpoint is created
Remote replication example:
To create a SnapSure checkpoint of source file system src_ufs1, type:
$ fs_ckpt src_ufs1 -Create
Local replication example:
To create a SnapSure checkpoint of source file system local_src, type:
$ fs_ckpt local_src -Create

Note: The following output shown is for the remote replication example.

50 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Output

operation in progress (not interruptible)...id = 88


name = src_ufs1
acl = 0
in_use = True
type = uxfs
worm = off
volume = v243
pool = clar_r5_performance
member_of = root_avm_fs_group_10
rw_servers= server_2
ro_servers=
rw_vdms =
ro_vdms =
ckpts = src_ufs1_ckpt1,src_ufs1_ckpt2
ip_copies = dst_ufs1:cs110
stor_devs = APM00034000068-001F,APM00034000068-001E
disks = d21,d15
disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14
server=server_2
disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13
server=server_2

id = 97
name = src_ufs1_ckpt2
acl = 0
in_use = True
type = ckpt
worm = off
volume = vp246
pool = clar_r5_performance
member_of =
rw_servers=
ro_servers= server_2
rw_vdms =
ro_vdms =
checkpt_of= src_ufs1 Mon Feb 7 07:05:00 EST 2005
used = 3%
full(mark)= 90%
stor_devs = APM00034000068-001F,APM00034000068-001E
disks = d21,d15
disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14
server=server_2
disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13
server=server_2

Using Celerra Replicator (V1) Version 5.6.47 51 of 172


Task 8: Copy incremental changes
To ensure that the file system type is uxfs when the copy completes, do not use the
convert=no option. Use of the monitor=off option runs this command as a
background process, which enables you to run several copy sessions
simultaneously.

Action

To copy the incremental changes (or the delta set) between the two source file system checkpoints
to the destination file system, use this command syntax:
$ fs_copy -start <new_check_point> <dstfs>:cel=<cel_name> -fromfs
<previous_check_point> -option monitor=off
where:
<new_check_point> = last checkpoint taken as described in Task 7: "Create a second
checkpoint of the source file system" on page 50.
<dstfs> = destination file system.
<cel_name> = Celerra Network Server where the destination file system resides.
<previous_check_point> = first checkpoint taken.
-option monitor=off = progress of the copy is printed to the screen by default. The off
option forces the command to run as a background process.

Remote replication example:


The command to use here differs, depending on how you copied your source file system to the
destination site. Use the -Force option only if you used a physical transport method (disk or tape).
To copy the incremental changes between checkpoints of source file system src_ufs1, to
destination file system dst_ufs1, type:
$ fs_copy -start src_ufs1_ckpt2 dst_ufs1:cel=cs110 -fromfs
src_ufs1_ckpt1 -option monitor=off
If you used a physical transport to perform the original source file system copy to the destination
site, type:
$ fs_copy -start src_ufs1_ckpt2 dst_ufs1:cel=cs110 -fromfs
src_ufs1_ckpt1 -Force -option monitor=off
Local replication example:
To copy the incremental changes between checkpoints of source file system local_src, type:
$ fs_copy -start local_src_ckpt2 local_dst -fromfs local_src_ckpt1 -
option monitor=off

Note: The following output shown is for the remote replication example.

52 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Output

operation in progress (not interruptible)...id = 97


name = src_ufs1_ckpt2
acl = 0
in_use = True
type = ckpt
worm = off
volume = vp246
pool = clar_r5_performance
member_of =
rw_servers=
ro_servers= server_2
rw_vdms =
ro_vdms =
checkpt_of= src_ufs1 Mon Feb 7 07:05:00 EST 2005
used = 3%
full(mark)= 90%
stor_devs = APM00034000068-001F,APM00034000068-001E
disks . = d21,d15
disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14
server=server_2
disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13
server=server_2

id = 126
name = dst_ufs1
acl = 0
in_use = True
type = rawfs
worm = off
volume = v272
pool = clar_r5_performance
member_of = root_avm_fs_group_3:cs110
rw_servers=
ro_servers= server_2
rw_vdms =
ro_vdms =
backup_of = src_ufs1 Mon Feb 7 07:05:00 EST 2005
stor_devs = APM00044603845-0008,APM00044603845-0007
disks = d8,d9
disk=d8 stor_dev=APM00044603845-0008 addr=c0t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c32t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c16t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c48t1l2
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c16t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c48t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c0t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c32t1l1
server=server_2

done

Using Celerra Replicator (V1) Version 5.6.47 53 of 172


Verify the copied incremental changes
Action

To verify that the fs_copy command completes, type:


$ fs_copy -list

Output

Local Source Filesystems


Id Source Destination Status %Remaining CommState

Local Destination Filesystems


Id Source Destination Status %Remaining CommState

Note

The fs_copy session is not listed in the output which indicates that the copy is complete.

Task 9: Verify file system conversion


At the destination site, verify file system conversion:

Action

To verify that the file system is converted to a uxfs type file system, type:
[destination_site]$ nas_fs -info dst_ufs1

54 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Output

id = 126
name = dst_ufs1
acl = 0
in_use = True
type = uxfs
worm = off
volume = v272
pool = clar_r5_performance
member_of = root_avm_fs_group_3:cs110
rw_servers=
ro_servers= server_2
rw_vdms =
ro_vdms =
backup_of = src_ufs1 Mon Feb 7 07:05:00 EST 2005
stor_devs = APM00044603845-0008,APM00044603845-0007
disks = d8,d9
disk=d8 stor_dev=APM00044603845-0008 addr=c0t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c32t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c16t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c48t1l2
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c16t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c48t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c0t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c32t1l1
server=server_2

Task 10: Check replication status


You can optionally check the status of all replication sessions that are running on a
Celerra system and an individual replication session:
◆ "List all replication sessions (optional)" on page 55
◆ "List individual replication session (optional)" on page 57
List all replication sessions (optional)
Action

To list the current active replication sessions, type:


$ fs_replicate -list

Using Celerra Replicator (V1) Version 5.6.47 55 of 172


Output

Source:
Local Source Filesystems
Id Source FlowCtrl State Destination FlowCtrl State Network
138 src_ufs1 inactive active dst_ufs1:cs110 inactive active alive

Local Destination Filesystems


Id Source FlowCtrl State Dest. FlowCtrl State Network

Destination:
Local Source Filesystems
Id Source FlowCtrl State Destination FlowCtrl State Network

Local Destination Filesystems


Id Source FlowCtrl State Dest. FlowCtrl State Network
165 src_ufs1:cs100 inactive active dst_ufs1 inactive active alive

Note

For local replication, the Destination FlowCtrl status and Network status always contain N/A (not
applicable).

56 of 172 Version 5.6.47 Using Celerra Replicator (V1)


List individual replication session (optional)
Execute this command from either Celerra Network Server for which the system
reports the replication status of each current delta set on source and destination
sites.

Action

To check the replication status and generate historical data about the replication up to the number
of lines specified, use this command syntax:
$ fs_replicate -info <fs_name> -verbose <number_of_lines>
where:
<fs_name> = name of the file system (in the example it is the source file system).
<number_of_lines> = lines to display historical replication data. The maximum number is 128.
Remote replication example:
To check the replication status of the replication relationship, type:
$ fs_replicate -info src_ufs1 -verbose 10
Local replication example:
To check the replication status of the replication relationship, type:
$ fs_replicate -info local_src -verbose 10

Note: The following output shown is for the remote replication example. "Appendix A: fs_replicate
-info output fields" on page 164 gives the output definition description for the fs_replicate -info
command.

Using Celerra Replicator (V1) Version 5.6.47 57 of 172


Output

id = 88
name = src_ufs1
fs_state = active
type = replication
replicator_state = active
source_policy = NoPolicy
high_water_mark = 600
time_out = 600
current_delta_set = 3
current_number_of_blocks = 1
flow_control = inactive
total_savevol_space = 1048576 KBytes
savevol_space_available = 917504 KBytes (Before Flow Control)

id = 126
name = dst_ufs1:cs110
type = playback
playback_state = active
high_water_mark = 600
time_out = 600
current_delta_set = 3
flow_control = inactive
total_savevol_space = 1048576 KBytes
savevol_space_available = 786432 KBytes (Before Flow Control)

outstanding delta sets: <None>

communication_state = alive
current_transfer_rate = ~ 13312 Kbits/second
avg_transfer_rate = ~ 169984 Kbits/second
source_ip = 192.168.168.18
source_port = 57273
destination_ip = 192.168.168.20
destination_port = 8888
QOS_bandwidth = 0 kbits/sec

| Source | Destination
Delta|Create Time Dur Blocks|Playback Time Dur Blocks DSinGroup
-----|-------------- ------ ------|-------------- ------ ------ --------
-
2 2005-02-08 00:20 1
1 2005-02-08 00:20 128 2005-02-08 00:20 128 2
0 2005-02-08 00:10 333 2005-02-08 00:10 333 1

Note

All times are GMT. Block size is 8 KB.

58 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Task 11: Create restartable checkpoints
To ensure that a valid checkpoint is available to restart an out-of-sync replication
relationship, create two checkpoints of the source file system. These restartable
checkpoints, named <source_fs_name>_repl_restart_1 and
<source_fs_name>_repl_restart_2, are automatically refreshed by a system CRON
job that runs every hour at 25 minutes after the hour. When a replication
relationship is out-of-sync, these checkpoints are available for the restart process.
You must include these two checkpoints when determining the maximum number of
checkpoints per file system used with replication.

Action

To create checkpoints for use when a replication relationship falls out-of-sync, use this command
syntax:
$ fs_ckpt <fs_name> -name <name> -Create
where:
<fs_name> = file system name on which a checkpoint is created
<name> = name of the refresh checkpoint that must follow this naming convention
<source_fs_name>_repl_restart_1 and <source_fs_name>_repl_restart_2
Example:
To create the checkpoints of the source file system for use when a replication relationship falls out-
of-sync, type:
$ fs_ckpt src_ufs1 -name src_ufs1_repl_restart_1 -Create
$ fs_ckpt src_ufs1 -name src_usf1_repl_restart_2 -Create
The output only shows the creation of the first checkpoint.

Using Celerra Replicator (V1) Version 5.6.47 59 of 172


Output

operation in progress (not interruptible)...id = 88


name = src_ufs1
acl = 0
in_use = True
type = uxfs
worm = off
volume = v243
pool = clar_r5_performance
member_of = root_avm_fs_group_10
rw_servers= server_2
ro_servers=
rw_vdms =
ro_vdms =
ckpts = src_ufs1_ckpt1,src_ufs1_ckpt2,src_ufs1_repl_restart_1
ip_copies = dst_ufs1:cs110
stor_devs = APM00034000068-001F,APM00034000068-001E
disks = d21,d15
disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14
server=server_2
disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13
server=server_2

id = 100
name = src_ufs1_repl_restart_1
acl = 0
in_use = True
type = ckpt
worm = off
volume = vp246
pool = clar_r5_performance
member_of =
rw_servers=
ro_servers= server_2
rw_vdms =
ro_vdms =
checkpt_of= src_ufs1 Mon Feb 7 07:14:26 EST 2005
used = 4%
full(mark)= 90%
delta_number= 1
stor_devs = APM00034000068-001F,APM00034000068-001E
disks = d21,d15
disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14
server=server_2
disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13
server=server_2

60 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Recovering replication data
Occasionally, you might need to recover from a situation in which a replication
relationship must be terminated because of a disaster situation. For example, the
source site experiences a disaster and becomes unavailable for data processing, or
the source file system becomes corrupted and is unrecoverable. In this situation:
1. You can perform a failover from the site where you replicate your data. A
checkpoint is created on the destination site which allows the system to track
changes that occur after the failover. The failover process makes the
destination file system read/write. "Replication failover" on page 62 provides
more information.
2. When the source site becomes available again, resynchronize your source and
destination sites and restart replication. However, after resynchronization
replication is running in the reverse direction. "Resynchronize the source and
destination sites" on page 66 provides more information.
3. Use a replication reversal to restore the direction of replication to what it was
prior to the failover. When you schedule a replication reversal, the write activity
on the destination file system is stopped, and the changes are applied to the
source file system before the source site becomes read/write. The reversal
process synchronizes the source and destination file systems. "Replication
reversal" on page 74 provides more information.

Note: Replication failover, resynchronization, and reversal must be performed sequentially.


For example, if you replicate multiple file systems from one Celerra system to another, you
must run separate failover commands and wait for one command to complete before
running the next command.

Table 7 on page 61 shows the relationships between the source and destination file
systems when there is a initiated replication, failover, resynchronization, and
reversal.
Table 7 Replication file system relationships (page 1 of 2)

Replication
Source site Destination site Explanation
option

Begin replication Source file Destination file system Normal replication processing
system is is read-only. establishes source file system
read/write. as read/write and destination
file system as read-only.

Failover Source file Destination file system Changes which file system
system becomes becomes read/write. acts as the source file system
read-only, if and which acts as the
source site is still destination file system.
available. Brings the destination file
system to read/write to service
the I/O in the case of disaster.

Using Celerra Replicator (V1) Version 5.6.47 61 of 172


Table 7 Replication file system relationships (page 2 of 2)

Replication
Source site Destination site Explanation
option

Resynchronize Source file Destination file system Use to reestablish replication


system remains remains read/write. after a failover.
read-only. Both sites must be available.

Reversal The read-only file The read/write file Changes which file system
system becomes system becomes read- acts as the source file system
read/write. only. and which acts as the
destination file system.
Perform a reversal from
whichever file system is
read/write.
When used after a failover,
restores the direction of
replication to what it was prior
to the failover.
Both sites must be available.

Task 1: Replication failover


Failover is the process that changes the destination file system from read-only to
read/write and stops the transmission of replicated data. The source file system, if
available, becomes read-only. Use the failover operation if the source site is
unavailable due to a disaster or if the source site is still available but you want to
activate the destination file system as read/write. Perform a failover using the
options specified in Table 8 on page 62.
Table 8 Failover options (page 1 of 2)

Failover Site that must


Use if What happens
option be available

default The source site is Plays back all available delta sets at the Destination site
totally corrupt or destination site before failing over.
unavailable.

now The source site is Initiates an immediate failover and no Destination site
totally corrupt or delta sets are played back.
unavailable.
Note: If you perform a failover using this
option and delta sets are in the SavVol
at the destination site, an incremental
resynchronization might not be possible
in all cases.

62 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Table 8 Failover options (page 2 of 2)

Failover Site that must


Use if What happens
option be available

sync The source site is Fails over without any data loss by Source and
still available. making the source file system read-only, destination
the destination file system read/write,
and creating a restart checkpoint of the
destination file system.
Synchronized failover takes longer to
invoke and cannot be performed if the
source site is unavailable. It is more
suited to a maintenance-related failover
as part of a failover plan.

Note: The sync option is not used in a


disaster situation because both sites
must be available.

Using the failover command when the source site is unavailable results in data loss
because delta sets cannot be transferred from the source site to the destination
site. Using the default failover option reduces the amount of data loss, by replaying
any available delta sets (pending data) at the destination before initiating failover.
Failover processing creates a file system checkpoint to use to resynchronize the
replication relationship. You must include this checkpoint when determining the
maximum number of checkpoints per file system used with replication.
When a failover completes, replication is stopped and the destination file system
becomes read/write. Replication is no longer running because the source site is
usually unavailable when a failover is initiated. If the source site becomes available,
reestablish replication in the opposite direction (from the destination site to the
source site) by resynchronizing the source and destination file systems.
For replication failover:
◆ "Verify status of destination file system" on page 64
◆ "Initiate replication failover" on page 64
◆ "Verify file system is read/write" on page 66

Using Celerra Replicator (V1) Version 5.6.47 63 of 172


Verify status of destination file system
Action

To check the status of the destination file system and verify if it is mounted as read-only, use this
command syntax:
$ server_mount <movername>
where:
<movername> = name of the Data Mover on which the file system is mounted
Example:
To verify the status of the destination file systems on server_2, type:
$ server_mount server_2

Note: The ro in the output indicates a read-only file system.

Output

server_2 :
root_fs_2 on / uxfs,perm,rw
fs1 on /fs1 uxfs,perm,rw
ckpt1 on /ckpt1 ckpt,perm,ro
fsk on /fsk ckpt,perm,ro
root_fs_common on /.etc_common uxfs,perm,ro
dst_ufsl on /dst_ufsl uxfs,perm,ro

Initiate replication failover


Action

To initiate a failover from the destination site, use this command syntax:
[destination_site]$ fs_replicate -failover <srcfs>:cel=<cel_name>
<dstfs> -option now|sync
where:
<srcfs> = source file system name
<cel_name> = Celerra Network Server name of the source site
<dstfs> = destination file system name
Example:
To fail over the source file system to the destination file system, type:
$ fs_replicate -failover src_ufs1:cel=cs100 dst_ufs1

64 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Output

operation in progress (not interruptible)...id = 126


name = dst_ufs1
acl = 0
in_use = True
type = uxfs
worm = off
volume = v272
pool = clar_r5_performance
member_of = root_avm_fs_group_3
rw_servers= server_2
ro_servers=
rw_vdms =
ro_vdms =
ckpts = root_restart_ckpt_88_2
stor_devs = APM00044603845-0008,APM00044603845-0007
disks = d8,d9
disk=d8 stor_dev=APM00044603845-0008 addr=c0t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c32t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c16t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c48t1l2
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c16t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c48t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c0t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c32t1l1
server=server_2

id = 88
name = src_ufs1
acl = 0
in_use = True
type = uxfs
worm = off
volume = v243
pool = clar_r5_performance
member_of = root_avm_fs_group_10:cs100
rw_servers= server_2
ro_servers=
rw_vdms =
ro_vdms =
ckpts = src_ufs1_repl_restart_1:cs100,src_ufs1_repl_restart_2:cs100
stor_devs = APM00034000068-001F,APM00034000068-001E
disks = d21,d15
disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14
server=server_2
disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13
server=server_2
done

Using Celerra Replicator (V1) Version 5.6.47 65 of 172


Verify file system is read/write
Action

To verify that the file system is mounted as read/write and is accessible to the network clients,
type:
$ server_mount server_2

Output

server_mount server_2
server_2 :
root_fs_2 on / uxfs,perm,rw
root_fs_common on /.etc_common uxfs,perm,ro
dst_ufs1 on /dst_ufs1 uxfs,perm,rw

Task 2: Resynchronize the source and destination sites


After the failover command is run, a checkpoint is created on the destination site
and the destination file system becomes read/write. New writes are then allowed on
the destination file system.
When replication is resynchronized, default values are used for the replication
policies. For example, high water mark and timeout are set to 600. You can specify
new policies when you restart replication using the any of these options:
◆ autofullcopy={yes}
◆ to=<timeout>
◆ dto=<destination timeout>
◆ hwm=<high_water_mark>
◆ dhwm=<destination high_water_mark>
◆ qos=<qos>
◆ autoro={yes|no}
◆ autofreeze={yes|no}

Note: If you need to increase your file system’s size and plan to resynchronize your source
and destination sites after a failover, you must complete the resynchronization (fs_replicate
-resync) before increasing the size of your destination file system.

To resynchronize the source and destination sites:


◆ "Verify file system is read-only" on page 67
◆ "Resynchronize the source and destination file systems and restart replication"
on page 68

66 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Verify file system is read-only

Step Action Result

1. Before resynchronizing the file server_2 :


systems, verify whether the file root_fs_2 on / uxfs,perm,rw
system on the original source root_fs_common on /.etc_common
uxfs,perm,ro
site is mounted as read-only by src_ufs1 on /src_ufs1 uxfs,perm,ro
typing:
$ server_mount server_2

Note: If the source file system


is not mounted read-only, the
following message
appears:Error 4124:
<file_system> : is not mounted
ro

2. If the file system (src_ufs1) in server_2 : done


the example is read/write,
change it to read-only by
typing:
$ server_mount server_2
-option ro src_ufs1

Using Celerra Replicator (V1) Version 5.6.47 67 of 172


Resynchronize the source and destination file systems and
restart replication
Action

To attempt to resynchronize the source and destination file systems and restart replication from
the destination site, use this command syntax:
[destination_site]$ fs_replicate -resync <dstfs>[:cel=<cel_name>]
<srcfs> -option autofullcopy=yes
where:
<dstfs> = current read-only file system name
<cel_name> = name of Celerra Network Server of the original source site
<srcfs> = current read/write file system name
-option autofullcopy=yes = executes a full copy of the file system if an incremental
resynchronization does not complete.
Example:
To resynchronize the file systems and resume replication, type:
$ fs_replicate -resync src_ufs1:cel=cs100 dst_ufs1
For the system to automatically perform a full copy of the source file system if the incremental
resynchronization fails, type:
$ fs_replicate -resync src_ufs1:cel=cs100 dst_ufs1 -option
autofullcopy=yes
The full copy of the file system using autofullcopy=yes can be time-consuming. Consider when
you want to run this command.

Note: If a disaster occurs during the transfer, some delta sets might become lost. As a result, the
replication process will not be able to completely replicate the source file system on the destination
site.

Output: Convert to rawfs

operation in progress (not interruptible)...Converting filesystem type


id = 88
name = src_ufs1
acl = 0
in_use = True
type = rawfs
worm = off
volume = v243
pool = clar_r5_performance
member_of = root_avm_fs_group_10
rw_servers=
ro_servers= server_2
rw_vdms =
ro_vdms =
ckpts = src_ufs1_repl_restart_2,src_ufs1_repl_restart_1
stor_devs = APM00034000068-001F,APM00034000068-001E
disks = d21,d15
disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14
server=server_2
disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13
server=server_2

68 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Output: Start copy

Starting baseline copy...


operation in progress (not interruptible)...id = 133
name = root_restart_ckpt_88_2
acl = 0
in_use = True
type = ckpt
worm = off
volume = vp284
pool = clar_r5_performance
member_of =
rw_servers=
ro_servers= server_2
rw_vdms =
ro_vdms =
checkpt_of= dst_ufs1 Mon Feb 7 07:25:11 EST 2005
used = 1%
full(mark)= 90%
delta_number= 4
stor_devs = APM00044603845-0008,APM00044603845-0007
disks = d8,d9
disk=d8 stor_dev=APM00044603845-0008 addr=c0t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c32t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c16t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c48t1l2
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c16t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c48t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c0t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c32t1l1
server=server_2

id = 88
name = src_ufs1
acl = 0
in_use = True
type = rawfs
worm = off
volume = v243
pool = clar_r5_performance
member_of = root_avm_fs_group_10:cs100
rw_servers=
ro_servers= server_2
rw_vdms =
ro_vdms =
backup_of = dst_ufs1 Mon Feb 7 07:25:11 EST 2005
ckpts = src_ufs1_repl_restart_2:cs100,src_ufs1_repl_restart_1:cs100
stor_devs = APM00034000068-001F,APM00034000068-001E
disks = d21,d15
disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14
server=server_2
disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13
server=server_2

IP Copy remaining (%) 100..Done.

done

Using Celerra Replicator (V1) Version 5.6.47 69 of 172


Output: Start copy

operation in progress (not interruptible)...id = 88


name = src_ufs1
acl = 0
in_use = True
type = rawfs
worm = off
volume = v243
pool = clar_r5_performance
member_of = root_avm_fs_group_10
rw_servers=
ro_servers= server_2
rw_vdms =
ro_vdms =
backup_of = dst_ufs1:cs110 Mon Feb 7 07:25:11 EST 2005
ckpts = src_ufs1_repl_restart_2,src_ufs1_repl_restart_1
stor_devs = APM00034000068-001F,APM00034000068-001E
disks = d21,d15
disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14
server=server_2
disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13
server=server_2
done

70 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Output: Starting replication

Starting replication...
operation in progress (not interruptible)...id = 126
name = dst_ufs1
acl = 0
in_use = True
type = uxfs
worm = off
volume = v272
pool = clar_r5_performance
member_of = root_avm_fs_group_3
rw_servers= server_2
ro_servers=
rw_vdms =
ro_vdms =
ckpts = root_restart_ckpt_88_2
ip_copies = src_ufs1:cs100
stor_devs = APM00044603845-0008,APM00044603845-0007
disks = d8,d9
disk=d8 stor_dev=APM00044603845-0008 addr=c0t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c32t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c16t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c48t1l2
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c16t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c48t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c0t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c32t1l1
server=server_2

id = 88
name = src_ufs1
acl = 0
in_use = True
type = rawfs
worm = off
volume = v243
pool = clar_r5_performance
member_of = root_avm_fs_group_10:cs100
rw_servers=
ro_servers= server_2
rw_vdms =
ro_vdms =
backup_of = dst_ufs1 Mon Feb 7 07:25:11 EST 2005
ckpts = src_ufs1_repl_restart_2:cs100,src_ufs1_repl_restart_1:cs100
stor_devs = APM00034000068-001F,APM00034000068-001E
disks = d21,d15
disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14
server=server_2
disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13
server=server_2
done

Using Celerra Replicator (V1) Version 5.6.47 71 of 172


Output: Generating a new checkpoint

Generating new checkpoint...


operation in progress (not interruptible)...id = 126
name = dst_ufs1
acl = 0
in_use = True
type = uxfs
worm = off
volume = v272
pool = clar_r5_performance
member_of = root_avm_fs_group_3
rw_servers= server_2
ro_servers=
rw_vdms =
ro_vdms =
ckpts = root_restart_ckpt_88_2,root_new_ckpt_dst_ufs1
ip_copies = src_ufs1:cs100
stor_devs = APM00044603845-0008,APM00044603845-0007
disks = d8,d9
disk=d8 stor_dev=APM00044603845-0008 addr=c0t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c32t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c16t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c48t1l2
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c16t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c48t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c0t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c32t1l1
server=server_2

id = 140
name = root_new_ckpt_dst_ufs1
acl = 0
in_use = True
type = ckpt
worm = off
volume = vp284
pool = clar_r5_performance
member_of =
rw_servers=
ro_servers= server_2
rw_vdms =
ro_vdms =
checkpt_of= dst_ufs1 Mon Feb 7 07:34:31 EST 2005
used = 3%
full(mark)= 90%
stor_devs = APM00044603845-0008,APM00044603845-0007
disks = d8,d9
disk=d8 stor_dev=APM00044603845-0008 addr=c0t1l2 server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c32t1l2 server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c16t1l2 server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c48t1l2 server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c16t1l1 server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c48t1l1 server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c0t1l1 server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c32t1l1 server=server_2

72 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Output: Starting differential copy

Starting diff copy...

operation in progress (not interruptible)...id = 140


name = root_new_ckpt_dst_ufs1
acl = 0
in_use = True
type = ckpt
worm = off
volume = vp284
pool = clar_r5_performance
member_of =
rw_servers=
ro_servers= server_2
rw_vdms =
ro_vdms =
checkpt_of= dst_ufs1 Mon Feb 7 07:34:31 EST 2005
used = 3%
full(mark)= 90%
stor_devs = APM00044603845-0008,APM00044603845-0007
disks = d8,d9
disk=d8 stor_dev=APM00044603845-0008 addr=c0t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c32t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c16t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c48t1l2
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c16t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c48t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c0t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c32t1l1
server=server_2

id = 88
name = src_ufs1
acl = 0
in_use = True
type = rawfs
worm = off
volume = v243
pool = clar_r5_performance
member_of = root_avm_fs_group_10:cs100
rw_servers=
ro_servers= server_2
rw_vdms =
ro_vdms =
backup_of = dst_ufs1 Mon Feb 7 07:34:31 EST 2005
ckpts = src_ufs1_repl_restart_2:cs100,src_ufs1_repl_restart_1:cs100
stor_devs = APM00034000068-001F,APM00034000068-001E
disks = d21,d15
disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14
server=server_2
disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13
server=server_2

IP Copy remaining (%) 100..Done.

done

Using Celerra Replicator (V1) Version 5.6.47 73 of 172


Output: Deleting checkpoints that are used to restart replication

Deleting root_restart_ckpt_88_2...
id = 133
name = root_restart_ckpt_88_2
acl = 0
in_use = False
type = ckpt
worm = off
volume =
rw_servers=
ro_servers=
rw_vdms =
ro_vdms =
Deleting root_new_ckpt_dst_ufs1...
id = 140
name = root_new_ckpt_dst_ufs1
acl = 0
in_use = False
type = ckpt
worm = off
volume =
rw_servers=
ro_servers=
rw_vdms =
ro_vdms =
Operation complete
done

Task 3: Replication reversal


A replication reversal is a scheduled change in the direction of replication and
requires both sites to be available. If you successfully resumed replication using the
-resync option, replication is now running in the reverse direction—from the
destination site to the source site. You can now execute a reversal to change the
direction of replication. This reversal returns the original source file system to
read/write and the original destination file system to read-only.
The write activity on the destination file system is stopped and any changes are
applied to the source file system before the source site becomes read/write. The
reversal process keeps the source and destination file systems synchronized. A
reversal is usually performed after a failover and resynchronization, but it can be
used at any time to reverse the replication direction.

Note: This can only be done when both sites are operational.

For replication reversal:


◆ "Verify direction of replication process" on page 75
◆ "Reverse the replication" on page 76

74 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Verify direction of replication process
Before you reverse the replication direction, verify which file system is read/write
and which is read-only by using the server_mount server_x command. There will be
a rw or ro entry to indicate the file system’s status. You then verify the direction.

Action

To verify the direction of your replication relationship, type:


$ fs_replicate -list

Note: Use this command on the site where the file system is read/write. In this example, the
destination file system is read/write because of the failover.

Output

fs_replicate -list
Local Source Filesystems
Id Source FlowCtrl State Destination FlowCtrl State Network
7 dst_ufs1 inactive active src_ufs1:cs100 inactive active alive

Local Destination Filesystems


Id Source FlowCtrl State Destination FlowCtrl State Network

Note

Notice the file system named dst_ufs1 is acting as the source file system (read/write) and the file
system named src_ufs1 is functioning as the destination file system (read-only).

Using Celerra Replicator (V1) Version 5.6.47 75 of 172


Reverse the replication
Action

To initiate a reversal from the original destination site, use this command syntax:
[original_destination_site]$ fs_replicate -reverse
<dstfs>:cel=<cel_name> <srcfs>
where:
<dstfs> = current read-only file system name. This file system will become the read/write file
system.
<cel_name> = name of Celerra Network Server at the original source site.
<srcfs> = current read/write file system name. This file system will become the read-only file
system.
Example:
To reverse the direction of the replication process and return the src_ufs1 file system to read/write
and the dst_ufs1 file system to read-only, type:
$ fs_replicate -reverse src_ufs1:cel=cs100 dst_ufs1

76 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Output

operation in progress (not interruptible)...id = 126


name = dst_ufs1
acl = 0
in_use = True
type = uxfs
worm = off
volume = v272
pool = clar_r5_performance
member_of = root_avm_fs_group_3
rw_servers= server_2
ro_servers=
rw_vdms =
ro_vdms =
ip_copies = src_ufs1:cs100
stor_devs = APM00044603845-0008,APM00044603845-0007
disks = d8,d9
disk=d8 stor_dev=APM00044603845-0008 addr=c0t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c32t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c16t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c48t1l2
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c16t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c48t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c0t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c32t1l1
server=server_2

id = 88
name = src_ufs1
acl = 0
in_use = True
type = uxfs
worm = off
volume = v243
pool = clar_r5_performance
member_of = root_avm_fs_group_10:cs100
rw_servers=
ro_servers= server_2
rw_vdms =
ro_vdms =
backup_of = dst_ufs1 Mon Feb 7 07:34:31 EST 2005
ckpts = src_ufs1_repl_restart_2:cs100,src_ufs1_repl_restart_1:cs100
stor_devs = APM00034000068-001F,APM00034000068-001E
disks = d21,d15
disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14
server=server_2
disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13
server=server_2
done

operation in progress (not interruptible)...


done

Using Celerra Replicator (V1) Version 5.6.47 77 of 172


Note

• After this command completes, the original source file system is read/write and the original
destination file system is read-only. Replication is now running in the direction it was before the
failover. You can verify this by using the server_mount server_x command.
• When replication is reversed, default values are used for the replication policies. For example,
high water mark and timeout are set to 600. You can specify new policies when you restart
replication using -option <options>.

After a failover or reversal


After you complete a failover or reversal, to enable users to access the file system
that has newly acquired read/write access permission, you must configure Celerra
system appropriately for NFS. In many cases, you can use the same configuration
you used for the source file system which might include:
◆ File system exports
◆ I18N mode (Unicode or ASCII) configuration
◆ Network interfaces
Replicating EMC Celerra CIFS Environments (V1) describes all CIFS replication
issues.

78 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Abort Celerra Replicator
Abort Celerra Replicator when you no longer want to replicate the file system or
when your source and destination file systems are not synchronized and you want
to end replication.
After Celerra Replicator successfully aborts, the destination file system is a read-
only file system at the destination site. The source file system is a read/write file
system at the source site. When a session has been aborted it cannot be restarted
using the
-restart option.

Note: Aborting replication does not delete the underlying file systems.

Note: Multiple fs_replicate -abort processes are executed sequentially, not in parallel. Only
run one fs_replicate -abort command at a time.

Action

To abort replication on source and destination file systems simultaneously, use this command
syntax from the source site:
$ fs_replicate -abort <srcfs>,<dstfs>:cel=<cel_name>
where:
<srcfs> = name of the source file system
<dstfs> = name of the destination file system
<cel_name> = name of the remote Celerra Network Server
Example:
To stop replication for the replication relationship, type:
[source_site]$ fs_replicate -abort src_ufs1,dst_ufs1:cel=cs110

Using Celerra Replicator (V1) Version 5.6.47 79 of 172


Output

operation in progress (not interruptible)...id = 88


name = src_ufs1
acl = 0
in_use = True
type = uxfs
worm = off
volume = v243
pool = clar_r5_performance
member_of = root_avm_fs_group_10
rw_servers= server_2
ro_servers=
rw_vdms =
ro_vdms =
ckpts = src_ufs1_repl_restart_2,src_ufs1_repl_restart_1
stor_devs = APM00034000068-001F,APM00034000068-001E
disks = d21,d15
disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14
server=server_2
disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13
server=server_2

id = 126
name = dst_ufs1
acl = 0
in_use = True
type = uxfs
worm = off
volume = v272
pool = clar_r5_performance
member_of = root_avm_fs_group_3:cs110
rw_servers=
ro_servers= server_2
rw_vdms =
ro_vdms =
stor_devs = APM00044603845-0008,APM00044603845-0007
disks = d8,d9
disk=d8 stor_dev=APM00044603845-0008 addr=c0t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c32t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c16t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c48t1l2
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c16t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c48t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c0t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c32t1l1
server=server_2
done

80 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Suspend a replication relationship
Suspend is an option that allows you to temporarily stop an active replication
relationship and leave replication in a condition that allows it to be restarted.

Action

To suspend replication, use this command syntax:


[source_site]$ fs_replicate -suspend <srcfs> <dstfs>:cel=<cel_name>
where:
<scrfs> = name of the source file system
<dstfs> = name of the destination file system
<cel_name> = name of the destination Celerra Network Server
Example:
To suspend a replication relationship, type:
$ fs_replicate -suspend src_ufs1 dst_ufs1:cel=cs110

Using Celerra Replicator (V1) Version 5.6.47 81 of 172


Suspend output: Creating a new delta set

operation in progress (not interruptible)...operation in progress (not


interruptible)...
id = 88
name = src_ufs1
fs_state = active
type = replication
replicator_state = active
source_policy = NoPolicy
high_water_mark = 0
time_out = 0
current_delta_set = 5
current_number_of_blocks = 0
flow_control = inactive
total_savevol_space = 1048576 KBytes
savevol_space_available = 917504 KBytes (Before Flow Control)

id = 126
name = dst_ufs1:cs110
type = playback
playback_state = active
high_water_mark = 600
time_out = 600
current_delta_set = 4
flow_control = inactive
total_savevol_space = 1048576 KBytes
savevol_space_available = 917504 KBytes (Before Flow Control)

outstanding delta sets:


Delta Source_create_time Blocks
----- ------------------ ------
4 2005-02-09 06:54:26 1

communication_state = alive
current_transfer_rate = ~ 13312 Kbits/second
avg_transfer_rate = ~ 13312 Kbits/second
source_ip = 192.168.168.18
source_port = 62817
destination_ip = 192.168.168.20
destination_port = 8888
QOS_bandwidth = 0 kbits/sec

Note: All times are in GMT. Block size is 8 KBytes.

done

82 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Suspend output: Creating a baseline checkpoint

Generating new checkpoint...


operation in progress (not interruptible)...id = 88
name = src_ufs1
acl = 0
in_use = True
type = uxfs
worm = off
volume = v243
pool = clar_r5_performance
member_of = root_avm_fs_group_10
rw_servers= server_2
ro_servers=
rw_vdms =
ro_vdms =
ckpts = src_ufs1_repl_restart_1,src_ufs1_ckpt1,
src_ufs1_repl_restart_2,root_susp end_ckpt_126_5
ip_copies = dst_ufs1:cs110
stor_devs = APM00034000068-001F,APM00034000068-001E
disks = d21,d15
disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14
server=server_2
disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13
server=server_2

id = 125
name = root_suspend_ckpt_126_5
acl = 0
in_use = True
type = ckpt
worm = off
volume = vp246
pool = clar_r5_performance
member_of =
rw_servers=
ro_servers= server_2
rw_vdms =
ro_vdms =
checkpt_of= src_ufs1 Tue Feb 8 13:51:01 EST 2005
used = 6%
full(mark)= 90%
delta_number= 5
stor_devs = APM00034000068-001F,APM00034000068-001E
disks = d21,d15
disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14
server=server_2
disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13
server=server_2

Using Celerra Replicator (V1) Version 5.6.47 83 of 172


Suspend output: Creating another delta set

operation in progress (not interruptible)...


id = 88
name = src_ufs1
fs_state = active
type = replication
replicator_state = active
source_policy = NoPolicy
high_water_mark = 0
time_out = 0
current_delta_set = 6
current_number_of_blocks = 0
flow_control = inactive
total_savevol_space = 1048576 KBytes
savevol_space_available = 917504 KBytes (Before Flow Control)

id = 126
name = dst_ufs1:cs110
type = playback
playback_state = active
high_water_mark = 600
time_out = 600
current_delta_set = 4
flow_control = inactive
total_savevol_space = 1048576 KBytes
savevol_space_available = 786432 KBytes (Before Flow Control)

outstanding delta sets:


Delta Source_create_time Blocks
----- ------------------ ------
5 2005-02-09 06:54:53 1
4 2005-02-09 06:54:26 1

communication_state = alive
current_transfer_rate = ~ 13312 Kbits/second
avg_transfer_rate = ~ 13312 Kbits/second
source_ip = 192.168.168.18
source_port = 62817
destination_ip = 192.168.168.20
destination_port = 8888
QOS_bandwidth = 0 kbits/sec

Note: All times are in GMT. Block size is 8 KBytes.

done

84 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Suspend output: Playing back delta set

operation in progress (not interruptible)...


id = 88
name = src_ufs1:cs100
fs_state = active
type = replication
replicator_state = active
source_policy = NoPolicy
high_water_mark = 0
time_out = 0
current_delta_set = 6
current_number_of_blocks = 0
flow_control = inactive
total_savevol_space = 1048576 KBytes
savevol_space_available = 917504 KBytes (Before Flow Control)

id = 126
name = dst_ufs1
type = playback
playback_state = active
high_water_mark = 0
time_out = 10
current_delta_set = 6
flow_control = inactive
total_savevol_space = 1048576 KBytes
savevol_space_available = 786432 KBytes (Before Flow Control)

outstanding delta sets: <None>

communication_state = alive
current_transfer_rate = ~ 13312 Kbits/second
avg_transfer_rate = ~ 13312 Kbits/second
source_ip = 192.168.168.18
source_port = 62817
destination_ip = 192.168.168.20
destination_port = 8888
QOS_bandwidth = 0 kbits/sec

Note: All times are in GMT. Block size is 8 KBytes.

done

Using Celerra Replicator (V1) Version 5.6.47 85 of 172


Suspend output: Waiting for synchronization

operation in progress (not interruptible)...id = 88


name = src_ufs1
acl = 0
in_use = True
type = uxfs
worm = off
volume = v243
pool = clar_r5_performance
member_of = root_avm_fs_group_10
rw_servers= server_2
ro_servers=
rw_vdms =
ro_vdms =
ckpts = src_ufs1_repl_restart_1,src_ufs1_ckpt1,
src_ufs1_repl_restart_2,root_susp end_ckpt_126_5
ip_copies = dst_ufs1:cs110
stor_devs = APM00034000068-001F,APM00034000068-001E
disks = d21,d15
disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14
server=server_2
disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13
server=server_2
done

86 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Suspend output: Convert destination file system to rawfs — ready for restart

name = dst_ufs1
acl = 0
in_use = True
type = uxfs
worm = off
volume = v272
pool = clar_r5_performance
member_of = root_avm_fs_group_3
rw_servers=
ro_servers= server_2
rw_vdms =
ro_vdms =
stor_devs = APM00044603845-0008,APM00044603845-0007
disks = d8,d9
disk=d8 stor_dev=APM00044603845-0008 addr=c0t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c32t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c16t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c48t1l2
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c16t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c48t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c0t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c32t1l1
server=server_2
done
Converting filesystem type
id = 126
name = dst_ufs1
acl = 0
in_use = True
type = rawfs
worm = off
volume = v272
pool = clar_r5_performance
member_of = root_avm_fs_group_3
rw_servers=
ro_servers= server_2
rw_vdms =
ro_vdms =
stor_devs = APM00044603845-0008,APM00044603845-0007
disks = d8,d9
disk=d8 stor_dev=APM00044603845-0008 addr=c0t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c32t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c16t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c48t1l2
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c16t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c48t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c0t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c32t1l1
server=server_2
Operation complete
done

Using Celerra Replicator (V1) Version 5.6.47 87 of 172


Verify the suspended replication relationship
Action

To verify that the replication relationship no longer exists, type either of the following two
commands. The -list option no longer displays the replication pair:
$ fs_replicate -list
$ fs_replicate -info src_ufs1

Output

Local Source Filesystems


Id Source FlowCtrl State Destination FlowCtrl State Network

Local Destination Filesystems


Id Source FlowCtrl State Dest. FlowCtrl State Network

The replication session is no longer listed.


Error 2242: src_ufs1 : replication/playback is not set up

88 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Restarting a replication relationship
To restart a replication relationship:
◆ "Verify that the replication relationship is not synchronized" on page 90
◆ "Restart replication relationship" on page 91

Using Celerra Replicator (V1) Version 5.6.47 89 of 172


Verify that the replication relationship is not synchronized
Action

To verify the replication relationship’s status, use this command syntax:


$ fs_replicate -info <srcfs>
where:
<srcfs> = name of the source file system
Example:
To verify the replication relationship, type:
$ fs_replicate -info src_ufs1 -verbose 10

Output

id = 88
name = src_ufs1
fs_state = active
type = replication
replicator_state = inactive
source_policy = NoPolicy
high_water_mark = 600
time_out = 600
current_delta_set = 0
current_number_of_blocks = 0
flow_control = active
total_savevol_space = 1048576 KBytes
savevol_space_available = 0 KBytes (Before Flow Control)

id = 126
name = dst_ufs1:cs110
type = playback
playback_state = active
high_water_mark = 600
time_out = 600
current_delta_set = 146
flow_control = inactive
total_savevol_space = 1048576 KBytes
savevol_space_available = 917504 KBytes (Before Flow Control)

outstanding delta sets: <None>

communication_state = down
current_transfer_rate = ~ 0 Kbits/second
avg_transfer_rate = ~ 0 Kbits/second
source_ip = 0.0.0.0
source_port = 0
destination_ip = 192.168.168.20
destination_port = 8888
QOS_bandwidth = 0 kbits/sec

| Source | Destination
Delta|Create Time Dur Blocks|Playback Time Dur Blocks DSinGroup
-----|-------------- ------ ------|-------------- ------ ------ --------
-

Note: All times are in GMT. Block size is 8 KBytes.

90 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Restart replication relationship
Action

To restart a replication relationship, use this command syntax from the source site:
$ fs_replicate -restart <srcfs> <dstfs>:cel=<cel_name>
where:
<scrfs> = name of the source file system
<dstfs> = name of the destination file system
<cel_name> = name of the destination Celerra Network Server
Example:
To restart a replication relationship, type:
$ fs_replicate -restart src_ufs1 dst_ufs1:cel=cs110

Note: The following output shown is for restarting an out-of-synchronization replication.

Using Celerra Replicator (V1) Version 5.6.47 91 of 172


Restart output: Converting file system type

operation in progress (not interruptible)...operation in progress (not


interruptible)...id = 88
name = src_ufs1
acl = 0
in_use = True
type = uxfs
worm = off
volume = v243
pool = clar_r5_performance
member_of = root_avm_fs_group_10
rw_servers= server_2
ro_servers=
rw_vdms =
ro_vdms =
ckpts = src_ufs1_repl_restart_2,src_ufs1_repl_restart_1
ip_copies = dst_ufs1:cs110
stor_devs = APM00034000068-001F,APM00034000068-001E
disks = d21,d15
disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14
server=server_2
disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13
server=server_2
done

operation in progress (not interruptible)...id = 126


name = dst_ufs1
acl = 0
in_use = True
type = uxfs
worm = off
volume = v272
pool = clar_r5_performance
member_of = root_avm_fs_group_3
rw_servers=
ro_servers= server_2
rw_vdms =
ro_vdms =
stor_devs = APM00044603845-0008,APM00044603845-0007
disks = d8,d9
disk=d8 stor_dev=APM00044603845-0008 addr=c0t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c32t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c16t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c48t1l2
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c16t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c48t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c0t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c32t1l1
server=server_2
done

92 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Restart output: Converting file system type

Converting filesystem type


id = 126
name = dst_ufs1
acl = 0
in_use = True
type = rawfs
worm = off
volume = v272
pool = clar_r5_performance
member_of = root_avm_fs_group_3
rw_servers=
ro_servers= server_2
rw_vdms =
ro_vdms =

Restart output: Converting file system type

stor_devs = APM00044603845-0008,APM00044603845-0007
disks = d8,d9
disk=d8 stor_dev=APM00044603845-0008 addr=c0t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c32t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c16t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c48t1l2
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c16t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c48t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c0t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c32t1l1
server=server_2

Using Celerra Replicator (V1) Version 5.6.47 93 of 172


Restart output: Starting replication

Starting replication...
operation in progress (not interruptible)...id = 88
name = src_ufs1
acl = 0
in_use = True
type = uxfs
worm = off
volume = v243
pool = clar_r5_performance
member_of = root_avm_fs_group_10
rw_servers= server_2
ro_servers=
rw_vdms =
ro_vdms =
ckpts = src_ufs1_repl_restart_2,src_ufs1_repl_restart_1
ip_copies = dst_ufs1:cs110
stor_devs = APM00034000068-001F,APM00034000068-001E
disks = d21,d15
disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14
server=server_2
disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13
server=server_2

id = 126
name = dst_ufs1
acl = 0
in_use = True
type = rawfs
worm = off
volume = v272
pool = clar_r5_performance
member_of = root_avm_fs_group_3:cs110
rw_servers=
ro_servers= server_2
rw_vdms =
ro_vdms =
backup_of = src_ufs1 Tue Feb 8 08:49:46 EST 2005
stor_devs = APM00044603845-0008,APM00044603845-0007
disks = d8,d9
disk=d8 stor_dev=APM00044603845-0008 addr=c0t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c32t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c16t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c48t1l2
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c16t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c48t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c0t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c32t1l1
server=server_2
done

94 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Restart output: Create checkpoint

Generating new checkpoint...


operation in progress (not interruptible)...id = 88
name = src_ufs1
acl = 0
in_use = True
type = uxfs
worm = off
volume = v243
pool = clar_r5_performance
member_of = root_avm_fs_group_10
rw_servers= server_2
ro_servers=
rw_vdms =
ro_vdms =
ckpts = src_ufs1_repl_restart_2,src_ufs1_repl_restart_1,
root_new_ckpt_src_ufs1
ip_copies = dst_ufs1:cs110
stor_devs = APM00034000068-001F,APM00034000068-001E
disks = d21,d15
disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14
server=server_2
disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13
server=server_2

id = 115
name = root_new_ckpt_src_ufs1
acl = 0
in_use = True
type = ckpt
worm = off
volume = vp246
pool = clar_r5_performance
member_of =
rw_servers=
ro_servers= server_2
rw_vdms =
ro_vdms =
checkpt_of= src_ufs1 Tue Feb 8 08:50:13 EST 2005
used = 4%
full(mark)= 90%
stor_devs = APM00034000068-001F,APM00034000068-001E
disks = d21,d15
disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14
server=server_2
disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13
server=server_2

Using Celerra Replicator (V1) Version 5.6.47 95 of 172


Restart output: Start differential copy

Starting diff copy...


operation in progress (not interruptible)...id = 115
name = root_new_ckpt_src_ufs1
acl = 0
in_use = True
type = ckpt
worm = off
volume = vp246
pool = clar_r5_performance
member_of =
rw_servers=
ro_servers= server_2
rw_vdms =
ro_vdms =
checkpt_of= src_ufs1 Tue Feb 8 08:50:13 EST 2005
used = 4%
full(mark)= 90%
stor_devs = APM00034000068-001F,APM00034000068-001E
disks = d21,d15
disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14
server=server_2
disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13
server=server_2

id = 126
name = dst_ufs1
acl = 0
in_use = True
type = rawfs
worm = off
volume = v272
pool = clar_r5_performance
member_of = root_avm_fs_group_3:cs110
rw_servers=
ro_servers= server_2
rw_vdms =
ro_vdms =
backup_of = src_ufs1 Tue Feb 8 08:50:13 EST 2005
stor_devs = APM00044603845-0008,APM00044603845-0007
disks = d8,d9
disk=d8 stor_dev=APM00044603845-0008 addr=c0t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c32t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c16t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c48t1l2
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c16t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c48t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c0t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c32t1l1
server=server_2

IP Copy remaining (%) 100..Done.

done

96 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Restart output: Delete restart checkpoints

Deleting root_new_ckpt_src_ufs1...
id = 115
name = root_new_ckpt_src_ufs1
acl = 0
in_use = False
type = ckpt
worm = off
volume =
rw_servers=
ro_servers=
rw_vdms =
ro_vdms =
Operation complete
done

Using Celerra Replicator (V1) Version 5.6.47 97 of 172


Extending the size of a file system
File systems can be extended automatically and manually. Follow these procedures
to extend file system size:
◆ "Extend file system size automatically" on page 98
◆ "Extend file system size" on page 101

Note: You cannot extend file systems using non-sliced volumes if replication is running.
When using Automatic File System Extension, the slice option must be enabled for an
Automatic Volume Management (AVM) pool when replication is running. Managing EMC
Celerra Volumes and File Systems with Automatic Volume Management describes how to
configure slices in detail.

Extend file system size automatically


File systems in a replication relationship can be extended automatically. For
replication, the Automatic File System Extension policy can be set on the source file
system only. When the policy is set, the destination file system is extended, and
then the source file system, just as with extending file system manually. This feature
is used on file systems created using AVM only. Managing EMC Celerra Volumes
and File Systems with Automatic Volume Management details Automatic File
System Extension.
Automatic File System Extension provides the following options in the CLI and in
Celerra Manager (they are called out here as they appear in Celerra Manager):
◆ Auto Extend enabled: To select the function.
◆ Virtual Provisioning™ enabled: To allocate storage capacity based on
anticipated need, but dedicate resources only as needed. When virtual
provisioning is enabled, the maximum file system size or real file system size,
whichever is larger is reported to NFS and CIFS clients through Celerra
Manager or CLI.
◆ High water mark: To specify the threshold, ranging from 50 to 90 percent of the
file system, which triggers file system extension.
◆ Maximum capacity (MB): The maximum size the file system can be extended.

Note: The virtual size of a source but not destination file system is visible from the
NFS/CIFS client.

98 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Action

To automatically extend the size of a file system (setting a high water mark, maximum size, and
virtual provisioning), use this command syntax:
$ nas_fs -modify <fs_name> -auto_extend yes -hwm <50-99>% -max_size
<integer>[T|G|M] -vp <yes|no>
where:
<fs_name> = name of the file system.
<50-99> = percentage of the file system that is full and in need of being reached before the file
system is automatically extended.
<integer> = file system’s maximum size (entered in TB, GB, or MB).
-vp = used with a specified Maximum Capacity value to report the anticipated or actual file system
size. When turned on, the virtual size is reported to clients.
Example:
To automatically extend the source file system src_ufs1 with a high water mark of 50% and a
minimum size of 70 MB with virtual provisioning enabled, type:
$ nas_fs -modify src_ufs1 -auto_extend yes -hwm 50% -max_size 70M -vp
yes

Output

id = 2707
name = pfs001
acl = 0
in_use = True
type = uxfs
worm = off
volume = v7283
pool = clarata_archive
member_of = root_avm_fs_group_10
rw_servers= 123secnfs
ro_servers=
rw_vdms =
ro_vdms =
auto_ext = hwm=50%,max_size=16777216M,virtual_provision=yes
ckpts = pfs001_ckpt60,pfs001_ckpt61,pfs001_ckpt62,
pfs001_ckpt63,pfs001_ckpt6
stor_devs = APM00043200225-0029,APM00043200225-002C,
APM00043200225-0027,APM00043
disks = d35,d21,d34,d20,d32,d16,d30,d14,d33,d17,d31,
d15,d38,d41,d36,d22,d29,d13,d39,d42,d28,d12,d37,d23
disk=d35 stor_dev=APM00043200225-0029 addr=c16t2l9
server=123secnfs
[nasadmin@lnsgc123 nasadmin]$

Using Celerra Replicator (V1) Version 5.6.47 99 of 172


Recover from Automatic File System Extension failure
Automatic File System Extension employs an internal script that checks for
adequate space on the source and destination sites. If replication is thwarted
because extension of either file system fails due to insufficient space, an error
message displays the cause. For example:
Nov 21 10:37:53 2005 CFS:3:101 fs auto extension failed: no space available to
extend src1
In the great majority of Automatic File System Extension failures, both file systems
fail to extend. By consulting the sys_log, you can determine whether only the
source file system extension failed if the log contains a line that directs you to use
the src_only option.
Whether one or both file system extensions fail, find the file systems’ sizes and
compare their block counts and volume sizes by using the nas_fs -size command.
The number displayed is a second way to determine which file system extension
failed. This method allows you to know how much space to reserve when you
manually extend the file systems. Perform the following steps to recover from a
failure to automatically extend the file system.

Step Action

1. Display the sys_log by typing:


$ cd /nas/log/
$ more sys_log

2. Find the size of the source and destination file system, and compare the values using this
command syntax:
$ nas_fs -size <fs_name>
The following example shows sample output:
total = 100837 avail = 99176 used = 1660 ( 1% ) (sizes in MB) (
blockcount = 209715200 )
volume: total = 102400 (sizes in MB) ( blockcount = 209715200 )

Note: Because total, available, and used values are generated from the operating system
and not updated until the destination file system is refreshed, that data will differ from total
volume data derived by the Control Station. Block # counts agree on the same file system,
however, and can be used to accurately compare totals on source and destination file
systems.

3. To extend the source file system, type the following command from the source site. For
size, type the difference between source and destination file systems determined in step 2:
$ nas_fs -xtend <fs_name> size=<integer>[T|G|M] -option src_only

Note: Because replication always extends the destination first, the source size will never
be larger than the destination size. If you specify the wrong size when you run this
command, the command will report the error and you must rerun the command with the
correct size.

100 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Extend file system size
To extend file system size:
◆ "Extend file system size manually" on page 101
◆ "Extend a file system after replication has been suspended" on page 103
◆ "Extend a file system after replication failover" on page 104
◆ "Start replication when the source file system is inactive" on page 104
Extend file system size manually
File systems can be manually extended while replication is running. If only the
destination file system extends, you can extend the source file system manually.
The procedure to do this (and extend both file systems) is the same for manual and
Automatic File System Extension. Extending the source file system extends the
destination file system by default. This maintains the same file system sizes at both
sites. "Recover from Automatic File System Extension failure" on page 100
provides more information.
Before extending the source file system, verify that the destination Celerra system
contains enough unused space. Use the nas_fs -size command to determine the
current size of the file system.

Action

To extend the source file system, use this command syntax:


[source_site]$ nas_fs -xtend <fs_name> size=<integer>[T|G|M] pool=<pool>
where:
<fs_name> = name of the source file system
<integer> = file system’s size in terabytes, gigabytes, or megabytes
<pool> = assigns a rule set for the file system
Example:
To extend the source file system src_ufs1 by 1024 megabytes, type:
$ nas_fs -xtend src_ufs1 size=1024M pool=clar_r5_performance

Using Celerra Replicator (V1) Version 5.6.47 101 of 172


Output

id = 88
name = src_ufs1
acl = 0
in_use = True
type = uxfs
worm = off
volume = v243
pool = clar_r5_performance
member_of = root_avm_fs_group_10
rw_servers= server_2
ro_servers=
rw_vdms =
ro_vdms =
ckpts = src_ufs1_repl_restart_2,src_ufs1_repl_restart_1
ip_copies = dst_ufs1:cs110
stor_devs = APM00034000068-001F,APM00034000068-001E
disks = d21,d15
disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14
server=server_2
disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13
server=server_2

Source file system size:


total = 3025 avail = 2010 used = 1014 ( 33% ) (sizes in MB) ( blockcount
= 6291456 )
volume: total = 3072 (sizes in MB) ( blockcount = 6291456 )

Destination file system size:


total = 3025 avail = 2010 used = 1014 ( 33% ) (sizes in MB) ( blockcount
= 6291456 )
volume: total = 3072 (sizes in MB) ( blockcount = 6291456 )

Note: Both file systems have the same block # count (6291456).

The nas_fs -xtend command extends the destination file system and then the source file system. If
either file system fails to extend, an error message displays the cause. If neither file system
extends, the following sample output appears:
Error 5008: Remote command failed:
remote celerra = cs0
remote exit status = 5
remote error = 0
remote message = CLSTD : volume(s) are not available

Note: Step 3 of "Recover from Automatic File System Extension failure" on page 100 explains how
to extend both file systems.

If only the destination file system extends, the following output appears:
Error 5008: pfsAvolume(s) are not available
PFS extension failed. Please extend PFS with “nas_fs -xtend... -option
src_only”

Note: Go to step 3 of the "Recover from Automatic File System Extension failure" on page 100 to
extend only the source file system.

102 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Extend a file system after replication has been suspended
If the source file system is extended automatically or manually after replication has
been suspended, perform the following steps to ensure that the source and
destination file systems are the same size before restarting replication.
If you do not perform these steps before restarting replication, the restart will fail
with an error because the source file system and destination file system are no
longer the same size (that is, the block counts differ). In this case, you must perform
these steps to recover from the error.

Step Action

1. Verify that the destination file system type is set to rawfs. If it is set to uxfs, convert the
destination file system from uxfs to rawfs by using this command syntax:
$ nas_fs -Type rawfs <dstfs> -Force
where:
<dstfs> = name of the destination file system
Example:
To verify that the file system is set to rawfs, type:
$ nas_fs -Type rawfs dst_ufs1 -Force

Note: A read-only file system must be set to rawfs prior to extending a file system or
restarting a replication.

2. Extend the destination file system manually using the same size as the source file system
by using this command syntax:
$ nas_fs -xtend <dst_fs> size=<integer>[T|G|M] -option <options>
where:
<dstfs> = name of the destination file system
<integer> = size of the secondary file system in terabytes, gigabytes, or megabytes
<option> = any comma-separated options, such as slice={y|n}, which specifies
whether the disk volumes used by the file system might be shared with other file systems
using a slice
Example:
To extend the destination file system dst_ufs1 by 2 MB, using the slice option, to match the
source file system extension to the same size, type:
$ nas_fs -xtend dst_ufs1 size=2M slice=y

3. Restart the replication relationship by using this command syntax from the source site:
$ fs_replicate -restart <srcfs> <dstfs>:cel=<cel_name>
where:
<scrfs> = name of the source file system
<dstfs> = name of the destination file system
<cel_name> = name of the destination Celerra Network Server
Example:
To restart a replication relationship, type:
$ fs_replicate -restart src_ufs1 dst_ufs1:cel=cs110

Using Celerra Replicator (V1) Version 5.6.47 103 of 172


Extend a file system after replication failover
After a replication failover ends, you must reconfigure the Automatic File System
Extension policy if you want to continue using the function on the destination file
system as your PFS. This is because the Celerra Network Server database, which
stores the automatic extension configuration, is not copied to the destination file
system during replication. As a result, the destination file system cannot be
extended automatically.
The procedure on page 98 provides an example, and Managing EMC Celerra
Volumes and File Systems with Automatic Volume Management offers detailed
instructions to configure this function’s policy.
If the source file system is still active and the original destination file system is
extended, and you want to resynchronize both file systems, follow this procedure:

Step Action

1. Verify that the source file system is mounted read-only. If it is not, do so now using this
command syntax:
$ server_mount <servername> -option ro <srcfs>/<srcfs_mountpoint>

2. Convert the source file system from uxfs to rawfs using this command syntax:
$ nas_fs -Type rawfs <srcfs> -Force

3. Manually extend the source file system and match the extended destination file system
size using this command syntax:
$ nas_fs -xtend <srcfs> size=<bytes>

4. Convert the source file system back to uxfs using this command syntax:
$ nas_fs -Type uxfs <srcfs> -Force

5. Resynchronize the destination and original source file systems using this command
syntax:
$ fs_replicate -resync <srcfs>:cel=<cel_name> <dstfs>

6. Reverse the replication and return to the original configuration using this command syntax:
$ fs_replicate -reverse <dstfs> <dstfs>:cel=<cel_name> <srcfs>

Start replication when the source file system is inactive


If the source file system is inactive, you have to consider other options. Consult
"Abort Celerra Replicator" on page 79 for guidance on whether to attempt an
incremental resynchronization of the source and destination file systems or start
replication from the beginning.

104 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Resetting replication policy
Replication policies are established when you initially set up the replication
relationship. Use the fs_replicate command to change the policies for each
replication session.
To reset replication policy:
◆ "High water mark and time-out policies" on page 105
◆ "Modify replication policy" on page 106
◆ "Change flow-control policies" on page 107
◆ "Set bandwidth size" on page 109
◆ "Set policies using parameters" on page 110

High water mark and time-out policies


High water mark and time-out policies prevent the replication service from creating
delta sets faster than it can copy them to the destination file system.
To avoid file systems dropping out-of-sync, reset these source file system policies
to higher values than the destination file system policies. The following example
resets the high water mark replication policy to:
◆ Trigger the replication service to create delta sets every 300 MB of change for
the source file system.
◆ Replay these delta sets to the destination file system every 300 MB of change.
Table 9 on page 105 explains the fs_replicate command options to change the
policies.
Table 9 Options to set high water mark and time-out policies

Option Behavior

-modify To specify values for the source and destination sites, include both file systems in
the command syntax. The values are effective the next time a trigger for these
policies is reached. For example, if policies are changed from 600 and 300 for high
water mark and time-out interval, respectively, the next time replication reaches
600, the trigger is changed to 300. If you set source high water mark and time-out
interval values without specifying values for the destination, the source values are
applied to the destination site.

-refresh A refresh replication initiates playback of outstanding data already on the


destination site and then creates a delta set of modified data on the source site.
This has the same effect as reaching the next high water mark or time-out interval.

Note: Multiple fs_replicate -refresh processes must be run sequentially, not concurrently.
Run only one fs_replicate -refresh command at a time.

Using Celerra Replicator (V1) Version 5.6.47 105 of 172


Modify replication policy
Action

To modify the replication policy for a source and destination file system, use this command syntax:
$ fs_replicate -modify <srcfs> -option hwm=<high_water_mark>,
to=<timeout>,dhwm=<high_water_mark>,dto=<timeout>
where:
<srcfs> = name of the source file system
-option hwm=<high_water_mark> = high water mark policy in megabytes
-option dto=<timeout> = time-out policy in seconds
Example:
To reset the high water mark for source and destination file systems, type:
$ fs_replicate -modify src_ufs1,dst_ufs1:cel=cs110 -option
hwm=300,dhwm=300

Output

operation in progress (not interruptible)...


id = 88
name = src_ufs1
fs_state = active
type = replication
replicator_state = active
source_policy = NoPolicy
high_water_mark = 600 (Pending: 300)
time_out = 600
current_delta_set = 11
current_number_of_blocks = 1
flow_control = inactive
total_savevol_space = 1048576 KBytes
savevol_space_available = 917504 KBytes (Before Flow Control)

id = 126
name = dst_ufs1:cs110
type = playback
playback_state = active
high_water_mark = 600 (Pending: 300)
time_out = 600
current_delta_set = 11
flow_control = inactive
total_savevol_space = 1048576 KBytes
savevol_space_available = 917504 KBytes (Before Flow Control)

outstanding delta sets: <None>

communication_state = alive
current_transfer_rate = ~ 13312 Kbits/second
avg_transfer_rate = ~ 13312 Kbits/second
source_ip = 172.24.168.123
source_port = 62815
destination_ip = 192.168.168.20
destination_port = 8888
QOS_bandwidth = 0 kbits/sec
Note: All times are in GMT. Block size is 8 KBytes.
done

106 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Note

Any changes to the time-out interval or high water mark occur when the next trigger point is
reached so the Pending entry shown above is removed when the policy value changes.

Change flow-control policies


When the source SavVol is full and the Data Mover can no longer track changes to
the source file system, use one of these two options to keep the replication session
active:
◆ Set a policy to freeze the file system, which temporarily halts all I/Os to the
source file system until sufficient space is available on the source SavVol.
◆ Set a policy that temporarily halts all writes to the source file system and makes
this file system read-only. This condition persists until sufficient space is
available on the source SavVol.
Consider the following when establishing these policies:
◆ These values take effect when the next trigger for creating a delta set is
reached.
◆ If replication is currently in one of these flow-control states, the new setting
takes effect after exiting the read-only or freeze situation.
◆ The autofreeze and autoro options can only be set on the source file system.
◆ Flow-control policies, by default, are not enabled, which can result in replication
becoming inactive. If that happens:
• Try to restart your replication relationship using the procedure described in
"Restarting a replication relationship" on page 89. If that is impossible, abort
(described in "Abort Celerra Replicator" on page 79) and restart replication
(described in "Initiating replication" on page 39).

Action

To change flow-control policies for a file system, use this command syntax:
$ fs_replicate -modify <fs_name>:cel=<cel_name> -option <options>
where:
<fs_name> = name of the source file system.
<cel_name> = name of the remote Celerra Network Server.
<options> = flow-control setting for the source file system. To freeze all I/O to the source file
system use autofreeze=yes. To allow users to continue read-only access to the source file
system use autoro=yes.
Example:
To freeze all I/O to the source file system, src_ufs1, specify the option autofreeze=yes. Type:
$ fs_replicate -modify src_ufs1 -option autofreeze=yes

Using Celerra Replicator (V1) Version 5.6.47 107 of 172


Output

operation in progress (not interruptible)...


id = 88
name = src_ufs1
fs_state = active
type = replication
replicator_state = active
source_policy = NoPolicy (Pending: Freeze)
high_water_mark = 600
time_out = 600
current_delta_set = 30
current_number_of_blocks = 0
flow_control = inactive
total_savevol_space = 1048576 KBytes
savevol_space_available = 917504 KBytes (Before Flow Control)

id = 126
name = dst_ufs1:cs110
type = playback
playback_state = active
high_water_mark = 600
time_out = 600
current_delta_set = 29
flow_control = inactive
total_savevol_space = 1048576 KBytes
savevol_space_available = 917504 KBytes (Before Flow Control)

outstanding delta sets:


Delta Source_create_time Blocks
----- ------------------ ------
29 2005-02-09 11:57:08 1

communication_state = alive
current_transfer_rate = ~ 13312 Kbits/second
avg_transfer_rate = ~ 13312 Kbits/second
source_ip = 192.168.168.18
source_port = 62819
destination_ip = 192.168.168.20
destination_port = 8888
QOS_bandwidth = 0 kbits/sec

Note: All times are in GMT. Block size is 8 KBytes.

done

Note

• Determine when the file system is in a read-only or freeze situation by using the fs_replicate -info
command and checking the fs_state field.
• When the file system is in a read-only situation, the fs_state field is romounted and when no I/O
is allowed, frozen.

108 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Set bandwidth size
Setting the bandwidth size, limits the total bandwidth for this replication session.
The default value is maximum available network bandwidth 0.

Action

To specify the maximum bandwidth used for a replication session, use this command syntax:
$ fs_replicate -modify <fs_name>:cel=<cel_name> -option qos=8000
where:
<fs_name> = name of the source file system
<cel_name> = name of the remote Celerra Network Server
<options> = qos = bandwidth typed in kilobytes per second
Example:
To set the maximum bandwidth for this replication session, type:
$ fs_replicate -modify src_ufs1 -option qos=8000

Note: This setting takes effect the next time data is sent across the IP network.

Using Celerra Replicator (V1) Version 5.6.47 109 of 172


Output

operation in progress (not interruptible)...


id = 88
name = src_ufs1
fs_state = active
type = replication
replicator_state = active
source_policy = NoPolicy
high_water_mark = 600
time_out = 600
current_delta_set = 30
current_number_of_blocks = 1
flow_control = inactive
total_savevol_space = 1048576 KBytes
savevol_space_available = 917504 KBytes (Before Flow Control)

id = 126
name = dst_ufs1:cs110
type = playback
playback_state = active
high_water_mark = 600
time_out = 600
current_delta_set = 30
flow_control = inactive
total_savevol_space = 1048576 KBytes
savevol_space_available = 917504 KBytes (Before Flow Control)

outstanding delta sets: <None>

communication_state = alive
current_transfer_rate = ~ 13312 Kbits/second
avg_transfer_rate = ~ 13312 Kbits/second
source_ip = 192.168.168.18
source_port = 62819
destination_ip = 192.168.168.20
destination_port = 8888
QOS_bandwidth = 8000 kbits/sec

Note: All times are in GMT. Block size is 8 KBytes.

done

Set policies using parameters


Use the fs_replicate command to set flow-control policies and bandwidth size for
each replication relationship rather than for all replication sessions on a Data
Mover. If you choose to set these policies for all sessions on a Data Mover, change
the parameter VRPL read-only and VRPL freeze for flow-control policies.
You can view and dynamically modify parameter values using the server_param
command or Celerra Manager graphical user interface. This technical module
describes only the command-line procedures. Celerra Manager online help details
how to use the graphical user interface to modify parameter values.

110 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Reverse the direction of a replication relationship
Use a replication reversal to change the direction of replication. You might want to
change the direction of replication to perform maintenance on the source site or to
do testing on the destination site.

Action

To change the direction of replication, use this command syntax:


[read/write side]$ fs_replicate -reverse <dstfs>:cel=<cel_name> <srcfs>
where:
<dstfs> = name of the file system currently read-only
<cel_name> = name of Celerra Network Server where the destination file system currently
resides
<srcfs> = name of the file system currently read/write
Example:
For the current read/write file system, src_ufs1, to become the read-only file system, type:
$ fs_replicate -reverse dst_ufs1:cel=cs110 src_ufs

Using Celerra Replicator (V1) Version 5.6.47 111 of 172


Output

operation in progress (not interruptible)...id = 88


name = src_ufs1
acl = 0
in_use = True
type = uxfs
worm = off
volume = v243
pool = clar_r5_performance
member_of = root_avm_fs_group_10
rw_servers= server_2
ro_servers=
rw_vdms =
ro_vdms =
ckpts =
src_ufs1_ckpt1,src_ufs1_repl_restart_2,src_ufs1_repl_restart_1
ip_copies = dst_ufs1:cs110
stor_devs = APM00034000068-001F,APM00034000068-001E
disks = d21,d15
disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14
server=server_2
disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13
server=server_2
disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13
server=server_2

id = 126
name = dst_ufs1
acl = 0
in_use = True
type = uxfs
worm = off
volume = v272
pool = clar_r5_performance
member_of = root_avm_fs_group_3:cs110
rw_servers=
ro_servers= server_2
rw_vdms =
ro_vdms =
backup_of = src_ufs1 Tue Feb 8 13:53:59 EST 2005
stor_devs = APM00044603845-0008,APM00044603845-0007
disks = d8,d9
disk=d8 stor_dev=APM00044603845-0008 addr=c0t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c32t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c16t1l2
server=server_2
disk=d8 stor_dev=APM00044603845-0008 addr=c48t1l2
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c16t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c48t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c0t1l1
server=server_2
disk=d9 stor_dev=APM00044603845-0007 addr=c32t1l1
server=server_2
done

112 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Note

• When this command completes, the current read/write file system (src_ufs1) becomes read-only
and the current read-only file system (dst_ufs1) becomes read/write.
• If you tried to run this command from the incorrect side (read-only), this error message appears:
Error 2247: this command must be issued on the current source site:cs100

Verify the reverse direction of replication relationship


Action

To verify if the direction of the replication reversed, type:


$ fs_replicate -list

Output

Local Source Filesystems


Id Source FlowCtrl State Destination FlowCtrl State
Network

Local Destination Filesystems


Id Source FlowCtrl State Dest. FlowCtrl State Network
135 dst_ufs1:cs110 inactive active src_ufs1 inactive active alive

Using Celerra Replicator (V1) Version 5.6.47 113 of 172


Monitor replication
Table 10 on page 114 shows the commands to use to monitor different aspects of
replication. You can also monitor replication using Celerra Manager, which is
described in Celerra Manager online help.
Table 10 Ways to monitor replication

Monitor Description Command

Data Movers Returns information about available server_sysstat <movername>


memory and the CPU idle percentage.

File systems Reports the amount of used and server_df <movername>


available disk space for a file system
and reports the amount of a file
system’s total capacity that is used.

Replication Shows information for all replication fs_replicate -list


sessions, as described in "List all
replication sessions (optional)" on
page 55.

Replication Shows information for an individual fs_replicate -info <fs_name>


replication session, as detailed in "List
individual replication session
(optional)" on page 57.

114 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Checking playback service and outstanding delta
sets
The -info option of fs_replicate lets you verify that the playback service is running
and which delta sets have not replayed to the destination file system. The following
tasks show how to determine that the playback service is running and to verify the
outstanding delta sets:
1. "Determine playback service status" on page 115
2. "Playback delta set" on page 118
3. "Verify delta set" on page 119

Task 1: Determine playback service status


Action

To determine the playback service is running, use this command syntax:


$ fs_replicate -info <fs_name> -verbose <number_of_lines>
where:
<fs_name> = file system name
<number_of_lines> = number of lines to display in the output
Example:
To determine the playback service is running, type:
$ fs_replicate -info src -verbose 20

Using Celerra Replicator (V1) Version 5.6.47 115 of 172


Output

$
id = 18
name = src
fs_state = active
type = replication
replicator_state = active
source_policy = NoPolicy
high_water_mark = 60000
time_out = 3600
current_delta_set = 92
current_number_of_blocks = 1
flow_control = inactive
total_savevol_space = 1048576 KBytes
savevol_space_available = 393216 KBytes (Before Flow Control)

id = 30
name = dest:eng168102
type = playback
playback_state = active
high_water_mark = 300
time_out = 600
current_delta_set = 87
flow_control = inactive
total_savevol_space = 1048576 KBytes
savevol_space_available = 393216 KBytes (Before Flow Control)

id = 30
name = dest:eng168102
type = playback
playback_state = active
high_water_mark = 300
time_out = 600
current_delta_set = 87
flow_control = inactive
total_savevol_space = 1048576 KBytes
savevol_space_available = 393216 KBytes (Before Flow Control)

outstanding delta sets:


Delta Source_create_time Blocks
----- ------------------ ------
90 06/04 10:08:54 1
89 06/04 09:58:54 1
88 06/04 09:58:43 1
87 06/04 09:48:43 14457
91 06/04 10:10:52 1

communication_state = alive
current_transfer_rate = ~ 13312 Kbits/second
avg_transfer_rate = ~ 18140.4 Kbits/second
source_ip = 10.168.0.11
source_port = 59068
destination_ip = 10.168.0.180
destination_port = 8888
QOS_bandwidth = 0 kbits/sec

116 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Output

| Source | Destination
Delta|Create Time Dur Blocks|Playback Time Dur Blocks DSinGroup
-----|-------------- ------ ------|-------------- ------ ------ --------
-
91 06/04 10:10:52 0 1
90 06/04 10:08:54 0 1
89 06/04 09:58:54 0 1
88 06/04 09:58:43 0 1
87 06/04 09:48:43 9 14457
86 06/04 09:39:25 0 1 06/04 09:44:35 0 1 1
85 06/04 09:29:25 0 1 06/04 09:34:35 0 1 1
84 06/04 09:19:25 0 1 06/04 09:24:35 0 1 1
83 06/04 09:09:25 0 1 06/04 09:14:35 0 1 1
82 06/04 08:59:25 0 1 06/04 09:04:34 0 1 1
81 06/04 08:49:25 0 1 06/04 08:54:34 0 1 1
80 06/04 08:39:25 0 1 06/04 08:44:34 0 1 1
79 06/04 08:29:25 0 1 06/04 08:34:34 0 1 1
78 06/04 08:19:25 0 1 06/04 08:24:34 0 1 1
77 06/04 08:09:25 0 1 06/04 08:14:34 0 1 1
76 06/04 07:59:25 0 1 06/04 08:04:34 0 1 1
75 06/04 07:49:25 0 1 06/04 07:54:34 0 1 1
74 06/04 07:39:25 0 1 06/04 07:44:34 0 1 1
73 06/04 07:29:25 0 1 06/04 07:34:34 0 1 1
72 06/04 07:19:25 0 1 06/04 07:24:34 0 1 1
Note: All times are in GMT. Block size is 8 KBytes.

Using Celerra Replicator (V1) Version 5.6.47 117 of 172


Task 2: Playback delta set
Action

To play back all delta sets up to a specified delta-set number, use this command syntax:
$ fs_replicate -refresh dest -option playuntildelta=91

Note: In this example, the system plays back all delta sets up to 91. Any delta sets greater than
that number will not replay.

Output

operation in progress (not interruptible)...


id = 18
name = src:eng16853
fs_state = active
type = replication
replicator_state = active
source_policy = NoPolicy
high_water_mark = 60000
time_out = 3600
current_delta_set = 92
current_number_of_blocks = 1
flow_control = inactive
total_savevol_space = 1048576 KBytes
savevol_space_available = 393216 KBytes (Before Flow Control)

id = 30
name = dest
type = playback
playback_state = active
high_water_mark = 300
time_out = 600
current_delta_set = 92
flow_control = inactive
total_savevol_space = 1048576 KBytes
savevol_space_available = 393216 KBytes (Before Flow Control)

outstanding delta sets: <None>


communication_state = alive
current_transfer_rate = ~ 13312 Kbits/second
avg_transfer_rate = ~ 18140.4 Kbits/second
source_ip = 10.168.0.11
source_port = 59068
destination_ip = 10.168.0.180
destination_port = 8888
QOS_bandwidth = 0 kbits/sec

Note: All times are in GMT. Block size is 8 KBytes.

done

118 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Task 3: Verify delta set
Action

To verify that the specified delta set was replayed, type:


$ fs_replicate -info src -verbose 20

Output

id = 18
name = src
fs_state = active
type = replication
replicator_state = active
source_policy = NoPolicy
high_water_mark = 60000
time_out = 3600
current_delta_set = 92
current_number_of_blocks = 1
flow_control = inactive
total_savevol_space = 1048576 KBytes
savevol_space_available = 393216 KBytes (Before Flow Control)

id = 30
name = dest:eng168102
type = playback
playback_state = active
high_water_mark = 300
time_out = 600
current_delta_set = 92
flow_control = inactive
total_savevol_space = 1048576 KBytes
savevol_space_available = 393216 KBytes (Before Flow Control)

outstanding delta sets: <None>

communication_state = alive
current_transfer_rate = ~ 13312 Kbits/second
avg_transfer_rate = ~ 18140.4 Kbits/second
source_ip = 10.168.0.11
source_port = 59068
destination_ip = 10.168.0.180
destination_port = 8888
QOS_bandwidth = 0 kbits/sec

| Source | Destination
Delta|Create Time Dur Blocks|Playback Time Dur Blocks DSinGroup
-----|-------------- ------ ------|-------------- ------ ------ --------
-
91 06/04 10:10:52 0 1 06/04 10:58:42 2 1 5
90 06/04 10:08:54 0 1 06/04 10:48:42 2 1 5
89 06/04 09:58:54 0 1 06/04 10:38:42 2 1 5
88 06/04 09:58:43 0 1 06/04 10:28:42 2 1 5
87 06/04 09:48:43 9 14457 06/04 10:18:42 2 14457 5
86 06/04 09:39:25 0 1 06/04 09:44:35 0 1 1
85 06/04 09:29:25 0 1 06/04 09:34:35 0 1 1
84 06/04 09:19:25 0 1 06/04 09:24:35 0 1 1
83 06/04 09:09:25 0 1 06/04 09:14:35 0 1 1
82 06/04 08:59:25 0 1 06/04 09:04:34 0 1 1
81 06/04 08:49:25 0 1 06/04 08:54:34 0 1 1
80 06/04 08:39:25 0 1 06/04 08:44:34 0 1 1

Using Celerra Replicator (V1) Version 5.6.47 119 of 172


Output

79 06/04 08:29:25 0 1 06/04 08:34:34 0 1 1


78 06/04 08:19:25 0 1 06/04 08:24:34 0 1 1
77 06/04 08:09:25 0 1 06/04 08:14:34 0 1 1
76 06/04 07:59:25 0 1 06/04 08:04:34 0 1 1
75 06/04 07:49:25 0 1 06/04 07:54:34 0 1 1
74 06/04 07:39:25 0 1 06/04 07:44:34 0 1 1
73 06/04 07:29:25 0 1 06/04 07:34:34 0 1 1
72 06/04 07:19:25 0 1 06/04 07:24:34 0 1 1
Note: All times are in GMT. Block size is 8 KBytes.

120 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Events for Celerra Replicator
Celerra Replicator provides the following file system events for use in SNMP traps
and email. System event traps and email notifications are configured by the user.
Configuring EMC Celerra Events and Notifications covers file system events.
Table 11 on page 121 explains Volume Replication (VRPL) facility events (Facility
ID : 77).
Table 11 Celerra Replicator events (page 1 of 2)

Event ID Event Description

0 Replication ok Not in use.

1 Replication on Source Filesystem Replication service is inactive.


Inactive

2 Resync asked by (previous) Request for a replication relationship


Destination Filesystem resynchronization after a failover.

3 Source Filesystem Switch Delta on Not in use.


HWM

4 Destination Filesystem in Error Playback service is inactive.

5 IP Rep Svc failed - Transport IPRepSender stopped.

6 IP Rep Svc NetWork or Receiver Down Source or destination IP network is down.

7 IP Rep Svc Network or Receiver Up Source or destination IP network is up.

8 Rep Svc Source Filesystem Flow on No source SavVol space available.


Hold

9 Rep Svc Source Filesystem Flow Source SavVol space available.


Resumed

10 Source Filesystem Frozen Result of specifying the autofreeze option or


parameter.

11 Source Filesystem Thawed Release of autofreeze option or parameter.

12 Last Delta Replayed on Destination Issued at the end of a failover and reverse.
Filesystem

13 Redo Buffer close to overflow event In write-intensive environments, sometimes


the redo buffers are not flushed fast enough.

14 Source mounted RO Result of specifying the autoro option or


parameter.

15 Source mounted RW Release of autoro option or parameter.

Using Celerra Replicator (V1) Version 5.6.47 121 of 172


Table 11 Celerra Replicator events (page 2 of 2)

Event ID Event Description

16 Replication in error Replication service is inactive or an internal


error occurred.

The events mentioned in Table 12 on page 122 apply to Celerra Replicator


functionality.
VMCAST : Events for fs_copy (Facility ID : 84)
Table 12 Events description

ID Description

0 Not an event

1 FS Copy over ip done

2 FS Copy over ip failed

3 Volume Copy over ip done

4 Volume Copy over ip failed

122 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Change the Celerra Replicator SavVol default size
By default, the system allocates 10 percent of the source file system size for each
Celerra Replicator SavVol. You can enlarge this default amount by changing the
value in the /nas/site/nas_param file.

Step Action

1. Log in to the Control Station.

2. Open /nas/site/nas_param with a text editor.


A short list of configuration lines appears.

! !
CAUTION
Do not edit anything in /nas/sys/, as these settings are overwritten with
each code upgrade.

3. Locate the replication configuration line:


Replication:10:
where:
10 = size (in %) of the source file system Celerra allocates for the replication SavVols.
The minimum SavVol size is 1 GB.
EMC recommends not using a value of less than 10%.

Note: Do not change any other lines in this file without a thorough knowledge of the
potential effects on the system. Contact EMC Customer Service for guidance.

4. Change the parameter to represent the percentage of space you want to allocate.

5. Save and close the file.

Note: Changing this value does not require a Control Station restart.

Using Celerra Replicator (V1) Version 5.6.47 123 of 172


Change the passphrase between Celerra Network
Servers
For a configuration with remote replication, use this procedure to change the
passphrase on the Celerra system at each site.

Step Action

1. At each site, you can review the current passphrase using this command syntax:
$ nas_cel -info <cel_name>
where:
<cel_name> = name of Celerra Network Server

2. At each site, establish the new passphrase using this command syntax:
$ nas_cel -modify <cel_name> -passphrase <passphrase>
where:
<cel_name> = name of Celerra Network Server
<passphrase> = new secure passphrase to be used for the connection, which must have
6- to 15-character and be the same on both sides of the connection
Example:
# nas_cel -modify cs110 -passphrase nas_replication
operation in progress (not interruptible)...
id = 5
name = cs110
owner = 503
device =
channel =
net_path = 192.168.168.102
celerra_id = APM000446038450000
passphrase = nas_replication

124 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Managing and avoiding IP replication problems
This section describes procedures to save replication sessions in the event of
network, and source or destination file system outages, planned or unplanned.
Additionally, this section details recommended practices to follow before performing
replication. Failure to follow these recommendations might lead to nasdb
inconsistencies and replication failures.
You can use Celerra Manager to perform much of the functionality described in this
section. Consult Celerra Manager online help for instructions.
This section describes:
◆ "Preventive measures to avoid IP replication problems" on page 125
◆ "Replication restart methods" on page 128
◆ "Recovering from a corrupted file system" on page 130
◆ "Recovering from an inactive replication state" on page 135
◆ "Managing anticipated destination site or network outages" on page 131
◆ "Managing unanticipated destination site or network outages" on page 132
◆ "Managing unanticipated source site outages" on page 133
◆ "Managing expected source site outages" on page 133
◆ "Mount the destination file system read/write temporarily" on page 133
◆ "Recovering from an inactive replication state" on page 135
◆ "Creating checkpoints on the destination site" on page 136
◆ "Copy file system to multiple destinations with fs_copy" on page 136

Preventive measures to avoid IP replication problems


You can take the following preventive measures to avoid IP replication problems:
◆ "Creating restartable checkpoints for out-of-sync operations" on page 125
◆ "Controlling delta set size" on page 126
◆ "Enlarging SavVol size" on page 126
◆ "Calculating modifications rate on the source file system" on page 127
◆ "Accommodating network concerns" on page 127
Creating restartable checkpoints for out-of-sync operations
In any IP replication recovery scenario, it is critical to have restartable checkpoints
as a base from which to work. Verifying that each file system contains checkpoints
that are occasionally refreshed ensures that Celerra can perform an out-of-sync
restart if required.
Restartable checkpoints created using fs_ckpt <srcfs> -name
<srcfs>_repl_restart_1 -Create and fs_ckpt <srcfs> -name <srcfs>_repl_restart_2 -
Create are automatically refreshed when starting from a differential copy.
Restartable checkpoints are supported in version 5.4 only.

Using Celerra Replicator (V1) Version 5.6.47 125 of 172


Be sure to verify that these checkpoints are being refreshed by using fs_ckpt
-list. For example, examine the creation times of the checkpoints, as shown in the
following sample output:
$ fs_ckpt pfs3 -list
id ckpt_name creation_time inuse full(mark) used
258 pfs3_ckpt1 01/03/2006-04:20:10-EST y 90% 10%
271 pfs3_ckpt2 01/03/2006-05:35:09-EST y 90% 10%
282 pfs3_ckpt3 01/03/2006-06:22:07-EST y 90% 10%
283 pfs3_ckpt4 01/03/2006-06:22:55-EST y 90% 10%
284 pfs3_ckpt5 01/03/2006-06:34:26-EST y 90% 10%

If restartable checkpoints do not exist, create them. If they do exist, but their
timestamps indicate they are not refreshing with the replication updates, check that
the names are correct and replication is healthy.
"Out-of-sync replication relationship" on page 20 provides more information about
these special checkpoints.
Controlling delta set size
Controlling the size of delta sets is integral to managing efficient replications.
Version 5.4 and later enforce an 8 GB delta-set limit. Manageable delta sets are
preferable to large deltas because:
◆ Large delta sets dramatically increase failover time (blockmap recovery).
◆ Large delta sets consume more operating system resources.
◆ Playback and create times for large delta sets do not rise proportionally with
size. Reasonably sized delta sets are created and replayed faster.
Enlarging SavVol size
Before beginning replication, determine whether the default source side SavVol
size (10 percent), as related to the file system size, is sufficient for the anticipated
delta sets.
Find the SavVol size by calculating 10 percent of the file system size obtained by
using nas_fs -size <fs_name>. If you previously changed the SavVol size, you can
learn its present size using the fs_replicate -info <fs_name> command and
checking the total_savevol_space field. To manage a large network outage or
account for brief intervals when the incoming modification rate significantly exceeds
the network’s ability to send changes to the destination site, you can increase the
size of the replication SavVol. For example, a 500 GB file system that incurs 20 GB
of change daily will, with a 50 GB SavVol, accommodate approximately two and
one-half days of outage. If replication has already begun, change the SavVol size
as follows:

Note: If you are starting replication, specify the SavVol size rather than use the default
value.

126 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Step Action

1. Record existing policy parameters using this command syntax:


$ fs_replicate -info <srcfs>
where:
<srcfs> = name of the source file system

2. Stop replication using this command syntax:


$ fs_replicate -suspend <srcfs> <dstfs>:cel=<cel_name>
where:
<srcfs> = name of the source file system
<dstfs> = name of the destination file system
<cel_name> = name of the Celerra Network Server for the file system

3. Restart replication with the revised SavVol and old parameters using this command
syntax:
$ fs_replicate -restart <srcfs> <dstfs>:cel=<cel_name>
savsize=<MB> -sav <srcsavvol_name> -option to=<value>, hwm=<MB>
where:
<srcfs> = name of the source file system
<dstfs> = name of the destination file system
<cel_name> = name of the Celerra Network Server for the file system
savsize=<MB> = size of the SavVol in MB
<srcsavvol_name> = name of the source file system SavVol
-option to=<value> = time-out interval in seconds
-option hwm=<value> = high water mark in MB

Calculating modifications rate on the source file system


Estimate the rate of modifications expected in the source side file system. If the rate
of anticipated change on the source file system is continuously greater than the
available network bandwidth, the replication service cannot transfer data quickly
enough to avoid becoming inactive. To avoid this state, create two restartable
checkpoints (consult "Start replication without initiating a full data copy" on
page 128). When autoro=yes the source file system becomes read only.
Accommodating network concerns
Consider the following quality of service (QoS) and network suggestions suitable to
your bandwidth requirements:
◆ Use dedicated network devices for IP replication data transfer to avoid an
impact on users.
◆ Apply QoS policy using a value that matches the bandwidth of one network to
another. For instance, even if you have a 1 MB/s line from A to B and want to fill
the pipe, set QoS at 1 MB/s so the line will not flood, causing packets to drop.
Configure the qos parameter to throttle bandwidth for a replication session (or
bandwidths for different sessions) on a Data Mover using fs_replicate <fsname>
-modify -option qos=<kbps>.
The EMC Celerra Network Server Command Reference Manual further details
the fs_replicate command.

Using Celerra Replicator (V1) Version 5.6.47 127 of 172


◆ If you do not expect significant packet loss, enable the fastRTO parameter.
fastRTO determines which TCP timer to use when calculating retransmission
timeout. The TCP slow timer (500 ms) is the default, causing the first time-out
retransmission to occur in 1–1.5 seconds. But setting fastRTO to 1 sets the TCP
fast timer (200 ms) for use when calculating retransmission timeout, which
causes the first timeout to occur in 400–600 ms.
Use server_param <movername> -facility tcp fastRTO=1 to configure the
setting.

Note: This setting may actually increase network traffic. Make the change cautiously,
recognizing it might not improve performance.

◆ To ensure that a stable network transfer rate for delta-set transfers on a Data
Mover, use a dedicated network port.
◆ Correctly set the TCP window size for network latency. Configuring the
tcpwindow parameter sets the window size used by replication (and fs_copy).
This value indicates the data load that can be sent before acknowledgment by
the receiving site. Increasing the value is most effective with a high latency.
Window size is calculated by multiplying the round-trip delay by the appropriate
packet rate.
For example, to send 10 MB/s across an IP network with a round-trip delay of
100 ms, a window size of 1 MB (0.1 sec x 10 MB/s = 1 MB) is needed. Use
server_param <movername> -facility rcp tcpwindow=bytes to configure the
setting.
EMC Celerra Network Server Parameters Guide provides more information on
setting tcp and rcp facilities.

Replication restart methods


If file systems are corrupted or out-of-sync, you can:
◆ "Start replication without initiating a full data copy" on page 128
◆ "Start replication from scratch" on page 129
You must start replication from scratch if the source or destination file system is
corrupted. If file systems are only out-of-sync, you can restart the replication
relationship.

Note: If you restart a replication session and there are unmounted checkpoints, a full data
copy will be initiated instead of a differential copy.

Start replication without initiating a full data copy


"Restarting a replication relationship" on page 89 provides details on how to start a
replication session without having to perform a full data copy.

128 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Start replication from scratch

Step Action

1. Terminate the replication relationship using this command syntax:


$ fs_replicate -abort <srcfs> <dstfs>:cel=<cel_name>
where:
<srcfs> = name of the source file system
<dstfs> = name of the destination file system
<cel_name> = name of Celerra Network Server for the file system

2. Create a checkpoint of the source file system using this command syntax:
$ fs_ckpt <srcfs> -Create
where:
<srcfs> = name of the source file system

3. Copy the checkpoint to the destination file system using this command syntax:
$ fs_copy -start <src_ckpt> <dstfs>:cel=<cel_name>
where:
<src_ckpt> = source checkpoint copied of the destination
<dstfs> = name of the destination file system
<cel_name> = name of the destination Celerra Network Server

4. Convert the destination file system to rawfs using this command syntax:
$ nas_fs -Type rawfs <dstfs> -Force
where:
<dstfs> = name of the destination file system

5. Start replication from the source to destination file system using this command syntax:
$ fs_replicate -start <srcfs> <dstfs>:cel=<cel_name>
where:
<srcfs> = name of the source file system
<dstfs> = name of the destination file system
<cel_name> = name of the destination Celerra Network Server

6. Create a second checkpoint of the source file system using this command syntax:
$ fs_ckpt <srcfs> -Create
where:
<srcfs> = name of the source file system

7. Perform a differential copy by typing:


$ fs_copy -start <src_newckpt> <dstfs>:cel=<cel_name> -fromfs
<previous_ckpt> monitor=off
where:
<dstfs> = name of the destination file system
<cel_name> = name of the destination Celerra Network Server

8. Check the copy’s progress and completion using this command syntax:
$ fs_copy -info <srcfs> or fs_copy -list
where:
<srcfs> = name of the source file system

Using Celerra Replicator (V1) Version 5.6.47 129 of 172


Recovering from a corrupted file system
To recover from a corrupted file system:
◆ "Run nas_fsck on the source file system" on page 130
◆ "Suspend and restart replication" on page 130

Note: You cannot use nas_fsck if the destination file system is corrupted as a result of an
improper fs_copy operation. File system replication fails due to pending nas_fsck.

Run nas_fsck on the source file system


The nas_fsck command checks and repairs a file system when replication is
threatened. You can run the command to check the source file system while
replication is running. If nas_fsck detects inconsistencies in the primary file system,
the changes that occur as a result of using nas_fsck are replicated to the
destination file system like any other file system modifications.

Step Action

1. Modify time-out and HWM values to zero using this command syntax:
$ fs_replicate -modify <srcfs> -option hwm=0,to=0
where:
<srcfs> = name of the source file system

Note: Setting trigger points to zero causes Celerra to keep replication active and track
changes, but not cut delta sets.

2. Run nas_fsck on the source file system to replicate and replay changes on the destination
file system using this command syntax:
$ nas_fsck -start <srcfs>
where:
<srcfs> = name of the source file system

Note: Running nas_fsck repairs corruption on the source file system, bringing it into a
consistent, but not original, state. While nas_fsck runs, the file system is not mounted to
avoid system instability. When the command is complete and inconsistencies addressed,
the file system is brought back online.

3. Revert to your previous time-out and HWM values using this command syntax:
$ fs_replicate -modify <srcfs> -option hwm=<MB>, to=<second>
where:
<srcfs> = name of the source file system
hwm=<value> = original high water mark value in MB
to=<value> = original time-out interval in seconds

Suspend and restart replication


If the destination is so unstable or corrupted that it might fail before the nas_fsck
changes are replicated and replayed, suspend replication, run nas_fsck -start
<srcfs>, and restart replication. While it is unlikely that only the destination file
system would be corrupted, you can diagnose a destination-only failure on the

130 of 172 Version 5.6.47 Using Celerra Replicator (V1)


source file system using the nas_fsck -start <srcfs> command. Any inconsistencies
found by nas_fsck on the primary file system are replicated to the secondary file
system.

Step Action

1. Suspend replication using this command syntax:


$ fs_replicate -suspend <srcfs> <dstfs>:cel=<cel_name>
where:
<srcfs> = name of the source file system
<dstfs> = name of the destination file system
<cel_name> = name of Celerra Network Server for the file system

2. Run nas_fsck using this command syntax at the source site:


$ nas_fsck -start <srcfs>
where:
<srcfs> = name of the source file system

3. Restart the replication using this command syntax:


$ fs_replicate -restart <srcfs> <dstfs>:cel=<cel_name>
where:
<srcfs> = name of the source file system
<dstfs> = name of the destination file system
<cel_name> = name of Celerra Network Server for the file system

Managing anticipated destination site or network outages


This section describes how to handle expected destination or network outages.
When planning for an outage, for instance, to restart a secondary Data Mover or to
conduct standard network maintenance involving the secondary file system, follow
the guidelines in the following sections to protect replication:
◆ "Anticipated destination site outage over a short period" on page 131
◆ "Anticipated destination site outage over a long period" on page 131
Anticipated destination site outage over a short period
Begin by evaluating the outage period and whether the site can survive it. If the
data queues easily in the SavVol, nothing need be done. For example, if the
planned outage period is one day, the SavVol is 100 MB, the file system is 1 GB,
and 200 MB of modifications occur daily, then survival is ensured for a half-day
because the SavVol will fill in 12 hours. On the other hand, if only 100 MB in
modifications occur daily, a whole day’s worth of changes are protected.
Anticipated destination site outage over a long period
In situations in which an outage is expected for a long period and the replication
service continues running, the SavVol might become full or trigger flow control,
eventually leading replication to drop out-of-sync.
If this scenario is likely, perform a replication suspend and restart, as directed in the
following procedure.

Using Celerra Replicator (V1) Version 5.6.47 131 of 172


Note: To restart after suspending replication, you must use the -restart option.

Step Action

1. Suspend each replication session individually using this command syntax:


$ fs_replicate -suspend <srcfs> <dstfs>:cel=<cel_name>
where:
<srcfs> = name of the source file system
<dstfs> = name of the destination file system
<cel_name> = name of the destination Celerra Network Server

Note: The suspend operation lets Celerra track changes in a SavVol that automatically
expands. Less total storage is needed because tracking is done only once in the
checkpoint SavVol, not twice in the checkpoint and replication SavVol. Also, checkpoints
retain only one overall changed block instead of one per delta set.

2. When all replication sessions are suspended, to check that the session you suspended no
longer appears in output for fs_replicate -list, type:
$ fs_replicate -list

Note: Use this command on source and destination Celerra systems to verify that no
sessions are running.

Output from the command should display no sessions, as shown below:


Local Source Filesystems
Id Source FlowCtrl State Destination FlowCtrl State Network

3. Verify the size of the suspend checkpoint to ensure that there is enough disk space to
expand the SavVol. The suspend checkpoint, root_suspend_ckpt, is used to restart
replication and is added by the -suspend option. Verify the size by typing:
$ nas_fs -size root_suspend_ckpt

4. Restart replication after the outage is over using this command syntax:
$ fs_replicate -restart <srcfs> <dstfs>:cel=<cel_name>
where:
<srcfs> = name of the source file system
<dstfs> = name of the destination file system
<cel_name> = name of the destination Celerra Network Server

Managing unanticipated destination site or network outages


This section describes coping with unanticipated Celerra destination platform or
network outages. Most destination outages require no action. For instance, if
Celerra restarts or goes offline—and the system does not fall out-of-sync—no
remediation is necessary. Consider the following:
◆ If fs_replicate -list shows replication is still active after the unplanned destination
outage is finished, nothing need be done.
◆ If fs_replicate -list shows replication is inactive and out-of-sync after the
unplanned destination outage is finished, consult "Replication restart methods"
on page 128.

132 of 172 Version 5.6.47 Using Celerra Replicator (V1)


◆ During an unplanned destination outage, the only way to mitigate the system
impact other than restoring the destination file system before the SavVol fills up
is to reconfigure a small timeout and HWM on Celerra, if applicable. In this
case, use fs_replicate -modify -option to=0, hwm=8192 to save a small amount
of SavVol space, allowing the system more time before falling out-of-sync. This
causes delta sets to be cut as large and with as little duplication as possible.
Contrast a TO of zero with a TO of 60 seconds, which consumes 128 MB of
SavVol per minute even without many modifications because the minimum
delta-set size is 128 MB.

Note: If replication failed and corrupted the destination file system, or you mistakenly
suspended replication, mounted the destination read/write, and restarted replication, you
must abort the session and restart using a full fs_copy. See "Replication restart methods" on
page 128 for more information.

Managing unanticipated source site outages


This section describes how to manage unplanned Celerra source outages such as
power interruptions or restarts.
Consider the following if you decide to activate the Data Recovery (DR) site:
◆ "Replication restart methods" on page 128 describes the source that resides on
a CLARiiON system that lost power, and replication is inactive.
◆ If the source resides on a Symmetrix system that lost power—where replication
should still be active—no remediation is necessary.
◆ If only the source is down and you do not want to activate the DR site, no
remediation is necessary.
◆ "Recovering replication data" on page 61 describes replication failover,
resynchronization, and reversal.

Managing expected source site outages


This section describes coping with both anticipated and unexpected outages on the
source Celerra. Perform the same steps as described in "Managing unanticipated
source site outages" on page 133, and decide whether to activate the DR site.

Mount the destination file system read/write temporarily


This section describes temporarily mounting the secondary file system as a
read/write volume to test data recovery without doing a failover and taking the
source server offline.
To temporarily mount the destination read/write, perform this procedure.

Using Celerra Replicator (V1) Version 5.6.47 133 of 172


! !
CAUTION
The following procedure cannot be used for VDM root file systems. A checkpoint of a
VDM’s root file system is designed in such a way that it cannot be restored.
Replication for VDM root file system needs to be restarted with a full fs_copy to
ensure that the source and destination file systems are synchronized. Failure to do
this will lead to file system corruption and data unavailability.

Step Action

1. Suspend replication using this command syntax:


$ fs_replicate -suspend <srcfs> <dstfs>:cel=<cel_name>
where:
<srcfs> = name of the source file system
<dstfs> = name of the destination file system
<cel_name> = name of Celerra Network Server for the file system

2. Assign the file system type to default using this command syntax:
$ nas_fs -Type uxfs <dstfs> -Force
where:
<dstfs> = name of the destination file system

3. Unmount the destination file system using this command syntax:


$ server_umount <movername> -perm <dstfs>
where:
<movername> = name of the Data Mover
<dstfs> = name of the destination file system

4. Mount the destination file system read-write using this command syntax:
$ server_mount <movername> -option rw <dstfs> /<dstfs_mountpoint>
where:
<movername> = name of the Data Mover
<dstfs> = name of the destination file system
<dstfs_mountpoint> = point at which the file system is mounted

5. Perform DR testing with an appropriate program after the both sites are read-write. A
number of diagnostics help to ensure that your database can start up correctly, such as:
• Reading/writing to a single file or every file
• Creating new files
• Modifying existing files
• Reading or deleting file systems

6. Make the now-writable destination file system available to clients either by exporting it
through NFS, or sharing it through CIFS (assuming a CIFS server is active on the
destination side) using this command syntax:
$ server_export <movername> -Protocol nfs -option <options>
<pathtoexport>
or
$ server_export <movername> -Protocol cifs -name <sharename> -
option <options> <pathtoshare>
where:
<movername> = name of the Data Mover

134 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Step Action

7. Unmount the destination file system after testing ends using this command syntax:
$ server_umount <movername> -perm <dstfs>
where:
<movername> = name of the Data Mover
<dstfs> = name of the destination file system

8. Mount the destination file system as read-only using this command syntax:
$ server_mount <movername> -option ro <dstfs> /<dstfs_mountpoint>
where:
<movername> = name of the Data Mover
<dstfs> = name of the destination file system
<dstfs_mountpoint> = point at which the file system is mounted

9. Convert the destination file system to rawfs after the restore ends using this command
syntax:
$ nas_fs -Type rawfs <dstfs> -Force
where:
<dstfs> = name of the destination file system

10. Restart replication at the source site using this command syntax:
$ fs_replicate -restart <srcfs> <dstfs>:cel=<cel_name>
where:
<srcfs> = name of the source file system
<dstfs> = name of the destination file system
<cel_name> = name of Celerra Network Server for the file system

Recovering from an inactive replication state


If the state of replication and playback is inactive due to a power failure on Celerra
or because flow control was exercised—and source and destination file systems
are out-of-sync—you must restart the replication relationship. Go to "Replication
restart methods" on page 128.

Using Celerra Replicator (V1) Version 5.6.47 135 of 172


Creating checkpoints on the destination site
When using the fs_copy command to regularly copy modifications from source to
destination file system, you should employ checkpoints on the destination site for
these reasons:
◆ While an fs_copy is in progress, the file system on the destination is in a rawfs
state and unavailable. However if you preserve a checkpoint on the destination,
this checkpoint remains available during the copy. "Copy file system to multiple
destinations with fs_copy" on page 136 provides more details.
◆ By using checkpoints, you can preserve the copied views of the file system on
the destination site. When checkpointing the destination, if an fs_copy is in
progress, the operation suspends until the copy is complete.
The following caveats apply when creating checkpoints on the destination file
system with replication running:
◆ Checkpoints created on the destination file system are supported just as on the
source file system, except that they fall on delta-set boundaries.
◆ Checkpoints processed during playback suspend until the playback ends. So, if
you are in the midst of a large delta-set playback, the refresh or create could
take an extended period to process while Celerra waits for the playback to
finish.

Copy file system to multiple destinations with fs_copy


The fs_copy command is used typically in a first-time replication start to manually
synchronize source and destination file systems. It is also often employed in a script
to copy one source file system to many destinations, either by cascading one file
system to a number of destinations or directly to each location.
Using fs_copy does not constitute replication. It is a copy command that emulates
replication which requires considerable scripting and integration. The command
should be used sparingly by a knowledgeable operator.
The following scenario describes regularly copying a file system to three destination
sites (A, B, and C). The procedure requires an operator to perform differential
copies on alternating checkpoint sets.

Note: It is highly recommended that slice volumes be used to create and extend file
systems for this fs_copy solution. Managing EMC Celerra Volumes and File Systems with
Automatic Volume Management provides more information about slice volumes.

136 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Step Action

1. Copy a file system to multiple destinations with fs_copy and create a checkpoint of the
source file system using this command syntax:
$ fs_ckpt <srcfs> -name <src_ckpt1> -Create
where:
<srcfs> = name of the source file system
<src_ckpt1> = first checkpoint on source

Note: If the primary file system extends before checkpoint creation, the fs_copy command
will fail. In this case, you must extend the destination file system manually to keep file
system sizes identical. First convert the file system to rawfs using the nas_fs -Type rawfs
command. Then use the nas_fs -xtend command. If a slice volume is not used, an
incremental copy might fail and you might need to run a full fs_copy from scratch when the
source file system is extended.

2. Copy it to the destination A file system after the checkpoint is created using this command
syntax:
$ fs_copy -start <src_ckpt1> <dstfs>:cel=<cel_name>
where:
<src_ckpt1> = first checkpoint on source
<dstfs> = destination name for file system A
<cel_name> = name of Celerra Network Server for file system A

Note: While an fs_copy is running, the destination file system is inaccessible. To make it
accessible, you must create a checkpoint before the data transfer starts.

3. After the fs_copy operation ends, create a checkpoint of destination file system using this
command syntax:
$ fs_ckpt <dstfs> -name <dst_ckpt1> -Create
where:
<dstfs> = destination file system name A
<dst_ckpt1> = first checkpoint on destination file system A

4. Create a second checkpoint of the source file system by typing:


$ fs_ckpt <srcfs> -name <src_ckpt2> -Create
where:
<srcfs> = name of the source file system
<src_ckpt2> = second checkpoint on source

5. Convert destination A file system to rawfs using this standard syntax:


$ nas_fs -Type rawfs <dstfs> -Force
where:
<dstfs> = destination name of file system A

Using Celerra Replicator (V1) Version 5.6.47 137 of 172


Step Action

6. Perform a differential copy between checkpoints 1 and 2 using this command syntax:
$ fs_copy -start <src_ckpt2> <dstfs>:cel=<cel_name> -fromfs
<src_ckpt1>
where:
<src_ckpt2> = second checkpoint on the source
<dstfs> = destination name for file system A
<cel_name> = name of Celerra Network Server for file system A
<src_ckpt1> = first checkpoint on the source

7. After the fs_copy command completes again, refresh the first checkpoint of destination file
system A using this command syntax:
$ fs_ckpt <dst_ckpt1> -refresh
where:
<dst_ckpt1> = first checkpoint on destination file system A

8. Perform steps 2 through 7 for destinations B and C. The source file system is now saved
on destinations A, B, and C.

9. Refresh the copy of the source file system at destinations A, B, and C, refresh the first
source checkpoint using this command syntax:
$ fs_ckpt <src_ckpt1> -refresh
where:
<src_ckpt1> = first checkpoint on source

10. Repeat steps 2 through 8, swapping checkpoints <src_ckpt1> and <src_ckpt2>.

138 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Transporting replication data using disk or tape
Copying the baseline source file system from the source to the destination site over
the IP network can be a time-consuming process. You can use an alternative
method by copying the initial checkpoint of the source file system, backing it up to a
disk array or tape drive, and transporting it to the destination site, as shown in
Figure 7 on page 139.

Source site Destination site

Celerra Data Celerra Data


Mover Mover

File File
system system

CLARiiON or CLARiiON or
Symmetrix Symmetrix
storage system storage system

Disk or Disk or
tape tape

Truck

CNS-000767

Figure 7 Physical transport of data

Note: Use the IP network to make the initial copy of the root file system for a VDM.

This section consists of the following tasks:


◆ "Disk transport method" on page 140
◆ "Tape transport method" on page 144

Using Celerra Replicator (V1) Version 5.6.47 139 of 172


Disk transport method
If the source file system holds a large amount of data, the initial copy of the source
to the destination file system can be time-consuming to move over the IP network.
Preferably, move the initial file system copy by disk, instead of over the network.
If you want to physically transport a copy of your source file system using a
CLARiiON disk array, the Data Movers that run replication must connect to the
storage system using Fibre Channel switched fabric connections.
To transport the baseline copy of the source file system to the destination site, use
the disk transport method described in this section or contact EMC about a
customized replication baselining service offering.

Note: You can use any qualified CLARiiON storage system for this transfer.

To transport replication data using disk:


1. "Capture data from the source site on disk" on page 141
2. "Transfer data to the destination site from disk" on page 143

140 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Step 1: Capture data from the source site on disk

Step Action

1. List the disks attached to the Celerra Network Server using the nas_disk -list command.
Keep this list to use for comparison later in this procedure.

2. Attach a supported CLARiiON array (for example a CX300) with the appropriately bound
LUNs to the Celerra Network Server.
This procedure assumes you will use a dedicated array. "Setting up the CLARiiON disk
array" on page 147 describes preparing the CLARiiON array.

3. Probe and verify SCSI disks by typing:


$ server_devconfig server_2 -probe -scsi -disks
where server_2 is the Data Mover with access to the CLARiiON array.

4. Create SCSI disks by typing:


$ nas_diskmark -mark -all
The disks will be available to all Data Movers including standby Data Mover.

5. List the disks that are attached to the Celerra Network Server using the nas_disk -list
command. Then perform a diff command between this list and the one created in step 1 of
"Capture data from the source site on disk" on page 141. For example:
> 378 n 260607 APM00034402893-0000 CLSTD d378 2
> 379 n 260607 APM00034402893-0001 CLSTD d379 2

6. Create a file system on the CLARiiON array that is the same size as the source file
system. To do so:
a. Create a user-defined pool by typing:
$ nas_pool -create -name transport_disks -volumes d378,d379
b. Create a file system by typing:
$ nas_fs -name transport_fs -type rawfs -create samesize=src
pool=transport_disks -option mover=server_2
c. Create a mountpoint for the file system.
d. Mount the file system read-only.

Note: Ensure that you create the file system as rawfs and use the samesize= option to
ensure that it is identical in size to the source file system.
When creating the pool, ensure that the disks are added in the same order on source and
destination sites. If you are creating more than one file system, ensure that they are
created in the same order on source and destination sites. EMC recommends that you
create the largest file system first.

7. Create a checkpoint of the source file system by typing:


$ fs_ckpt src -Create

8. Copy the source file system checkpoint to the file system created on the new disks by
typing:
$ fs_copy -start src_ckpt1 transport_fs -option convert=no,
monitor=off

Using Celerra Replicator (V1) Version 5.6.47 141 of 172


Step Action

9. Monitor the fs_copy progress by typing:


$ fs_copy -list
$ fs_copy -info <session_id>
Verify the size of the transport file system to the source file system using the
nas_fs -size command.

10. Delete the disks from the source site. To do so:


a. Unmount the file system transport_fs (server_umount)
b. Delete the mountpoint (server_mountpoint)
c. Delete the file system (nas_fs -delete)
d. Delete the pool (nas_pool -delete)
e. Delete the disks which were the result of the -diff command in step 5 (nas_disk -delete)
of "Capture data from the source site on disk" on page 141

11. Verify that the disks were removed by using the nas_disk -list command. The results you
obtain from this step should be the same as those derived from the first step of "Capture
data from the source site on disk" on page 141.

12. Disconnect and uninstall the CLARiiON array from the source site.

13. Transport the disk array to the destination site.

142 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Step 2: Transfer data to the destination site from disk

Step Action

1. List the disks that are attached to the Celerra Network Server using the nas_disk -list
command. Keep this list to use for comparison later in this procedure.

2. Attach the CLARiiON array at the destination site.

3. Probe and verify SCSI disks by typing:


$ server_devconfig server_2 -probe -scsi -disks
where server_2 is the Data Mover with access to the CLARiiON array.

4. Create SCSI disks by typing:


$ nas_diskmark -mark -all
The disks will be available to all Data Movers including standby Data Mover.

5. List the disks that are attached to the Celerra Network Server using the nas_disk -list
command. Then perform a diff command between this list and the one created in step 1 of
"Transfer data to the destination site from disk" on page 143. For example:
> 375 n 260607 APM00034402893-0000 CLSTD d375 1
> 376 n 260607 APM00034402893-0001 CLSTD d376 1

6. Create a file system on the transport disk array. To do so:


a. Create a user-defined pool by typing:
$ nas_pool -create -name transport_disks -volumes d375,d376
b. Create a file system by typing:
$ nas_fs -name transport_fs -type rawfs -create
samesize=src:cel=eng16853 pool=transport_disks -option
mover=server_2
c. Create a mountpoint.
d. Mount the file system read-only.

Note: Ensure that you create the file system as rawfs and use the samesize= option to
ensure that it is identical in size to the source file system. When creating the pool ensure
that the disks are added in the same order on source and destination sites. If you are
creating more than one file system ensure that they are created in the same order on
source and destination sites.

7. Create a destination file system (in this example the destination Celerra file system is
attached to a Symmetrix storage system). To do so:
a. Create a file system by typing:
$ nas_fs -name dest -type rawfs -create samesize=src:cel=eng16853
pool=symm_std
b. Create a mountpoint and mount the file system.

Note: Ensure that you create the file system as rawfs and use the samesize= option to
ensure that it is identical in size to the source file system.

8. Copy the file system on the transport disk array to the destination file system, created in
step 7 of "Transfer data to the destination site from disk" on page 143:
$ fs_copy -start transport_fs dest -option convert=no

Using Celerra Replicator (V1) Version 5.6.47 143 of 172


Step Action

9. The destination file system is now rawfs and contains a copy of the source file system
checkpoint.

10. Delete the disks from the transport disk array. To do so:
a. Unmount the file system transport_fs (server_umount)
b. Delete the mountpoint (server_mountpoint)
c. Delete the file system (nas_fs -delete)
d. Delete the pool (nas_pool -delete)
e. Delete the disks which were the result of the -diff command in step 5 (nas_disk -delete)
of "Transfer data to the destination site from disk" on page 143.

11. Verify that the disks were removed by using the nas_disk -list command. The results you
obtain from this step should be the same as those derived from step 1 of "Transfer data to
the destination site from disk" on page 143.

12. Disconnect and uninstall the CLARiiON array from the destination site.

13. Continue with the next step for setting up replication, Task 6: "Begin replication" on
page 48, in Using EMC Celerra Replicator (V1).

Tape transport method


If the source file system contains a large amount of data, the initial copy of the
source file system to the destination file system can be time-consuming to move
over the IP network. Moving the initial copy of the file system by backing it up to
tape, instead of over the network, is preferable.
When using this method of transport, note the following:
◆ You must have a valid NDMP infrastructure on both Celerra Network Servers.
◆ The restore is performed to a rawfs file system that must be mounted on the
Data Mover.
◆ The restore will be rejected if the destination file system is not the same size as
the source file system used for the backup.

Note: This special backup is used only for transporting replication data.

! !
CAUTION
Backing up file systems from a Unicode-enabled Data Mover and restoring to an
ASCII-enabled Data Mover is not supported. I18N mode (Unicode or ASCII) must be
the same on the source and destination Data Movers.

To transport replication data using tape:


1. "Capture data from the source site on tape" on page 145
2. "Transfer data to the destination site from tape" on page 146

144 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Step 1: Capture data from the source site on tape

Step Action

1. Create the checkpoint of the source file system by typing:


$ fs_ckpt src_ufs1 -Create

Note: You can back up only the checkpoint of an IP replication read-only target file system
using the NDMP backup feature. If you attempt to back up a replication read-only target
file system, NDMP will fail when replication is updating the change. Celerra Network
Server version 5.5.27 and later supports NDMP backup of integrated checkpoints and
manually created checkpoints of a target replication file system.

2. Set the NDMP environment variable for your backup software. For example, set the
VLC=y NDMP environmental variable before you run the backup.
The NDMP technical module for your particular backup software provides information
about environment variables. For information about how to set this variable, read your
backup software vendor’s documentation.

Note: The source file system and the checkpoint must be mounted on the NDMP Data
Mover.

3. Use your normal backup procedure to back up the source file system checkpoint.

4. Transport the backup tapes to the destination site.

Using Celerra Replicator (V1) Version 5.6.47 145 of 172


Step 2: Transfer data to the destination site from tape

Step Action

1. When the tapes are on the destination Celerra Network Server, create a file system (rawfs)
that is the same size as the source file system. Create the file system on a metavolume,
create a mount point, and then mount the file system. Managing EMC Celerra Volumes
and File Systems Manually describes how to create a file system.
Ensure that you create the file system as rawfs and use the samesize= option to ensure
that it is identical in size to the source file system.

2. Determine the volume number of the destination file system created in step 1. In this
example, the volume number for the rawfs file system is 66:
$ nas_fs -list
id inuse type acl volume name server
1 y 1 0 66 rawfs server_2
2 y 1 0 68 new server_3
3 y 1 0 70 last server_4

3. Using your normal NDMP restore procedure, restore the backup using the following as the
file system name:
/.celerra_vol_<fs_volume_ID>
where:
<fs_volume_ID> = volume number of the rawfs file system (66 in the example)

Note: The file system must be restored to the NDMP Data Mover.

4. The destination file system is now rawfs and contains the source file system checkpoint.
Start the replication between source and destination. Follow the procedure, Task 6: "Begin
replication" on page 48, to set up remote replication.

5. Create a second checkpoint. Follow the procedure, Task 7: "Create a second checkpoint
of the source file system" on page 50.

6. Perform an incremental copy and allow the destination system to convert to uxfs. Follow
the procedure, Task 8: "Copy incremental changes" on page 52.
Be sure to specify the -force option, for example:
$ fs_copy -start src_ufs1_ckpt2 dst_ufs1:cel=cs110 -fromfs
src_ufs1_ckpt1 -Force -option monitor=off

7. Verify the replication session:


$ fs_replicate -list

146 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Setting up the CLARiiON disk array
If you want to physically transport a copy of your source file system using a
CLARiiON disk array, you must set up replication on a Celerra gateway.
Use the following documents in this setup procedure:
◆ EMC CLARiiON CX300, CX500, and CX700 Storage Systems Initialization
Guide
◆ EMC CLARiiON CX300 2-Gigabit Disk Processor Enclosure (DPE2) Setup and
Cabling Guide

Note: Use the appropriate setup and cabling guide depending on the disk array used.

When configuring the CLARiiON disk array for transporting replication data, run the
appropriate setup script after setting up zoning for the network switches.
To prepare the CLARiiON disk array to receive the copy of the source file system:
◆ "Review the prerequisites" on page 147
◆ "Run the setup script" on page 149
◆ "Create data LUNs" on page 151

Review the prerequisites

Step Action

1. Cable and zone the CXxxx disk array to the Celerra Network Server.

2. Ensure that the required software components are installed on the CLARiiON disk array:
• CXxxx Base Array (EMC FLARE®)
• EMC Navisphere®_ArrayAgent
• Navisphere Management UI
• EMC Access Logix™
Read the E-Lab Interoperability Navigator for the most recent Celerra software and FLARE
microcode compatibility specifications.

3. The E-Lab Interoperability Navigator, which is available at Powerlink as definitive


information on supported software and hardware, such as backup software, Fibre Channel
switches, and application support for Celerra network-attached storage (Celerra) products.
The E-Lab Interoperability Navigator is for EMC use only. Do not share this information
with customers.

Using Celerra Replicator (V1) Version 5.6.47 147 of 172


Step Action

4. Make sure you have the following:


• Network information (hostname, IP address, subnet mask, and gateway for the SPs).
• Set up the service computer with:
• Windows Server 2000 or Windows NT 4.0 with Service Pack 5 (or later).
• Dial-up networking using direct connection.
• The latest version of the Navisphere CLI is installed in C:\program files\emc\navisphere
cli on the service computer.
• Null modem cable with 9-pin female-to-female connectors.
• An IP connection between the Celerra Network Server and the SPs (required for
installing the system software).

5. Create the PPP link by adding a modem and creating a connection as described in the
EMC CLARiiON CX300, CX500, and CX700 Storage Systems Initialization Guide.

148 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Run the setup script

Step Action

1. Using the NULL modem cable, connect the service computer to the SP A serial port.

Serial port

CNS-000760

2. Establish a dial-up connection between the service computer and SP A by selecting,


Start > Settings > Network and Dial-up Connections > Direct Connection.
The Connect Direct Connection dialog box opens.

3. Establish a dial-up connection by filling in the dialog box as follows:


Username: clariion
Password: clariion!

4. Select Connect.

Note: Establishing this connection might require several redial attempts.

5. Open a Command Prompt Window by selecting Start > Programs > Accessories >
Command Prompt.
The Command Prompt window opens.

6. Insert the Celerra Installation CD into the CD-ROM drive.

7. In the Command Prompt window, change to the \clariion directory on the CD-ROM by
typing the drive letter and then typing cd \clariion.
For example, type:
D:cd\clariion

Using Celerra Replicator (V1) Version 5.6.47 149 of 172


Step Action

8. Run the script by typing setup, and press Enter:


===========================================================
Program commandline: setup
Startup values:
Version: 7.0
Jumppoint into program:
Pathname to navicli: “c:\program files\emc\navisphere
cli\navicli”
Debug (1=yes) :
Logging (1=yes): 1
Logfile:
C:\DOCUME~1\ADMINI~1\LOCALS~1\Temp\laptop_script.log
IP Address for PPP: 128.221.252.1
Commands will all be sent via the dial-up connection.

Disconnect any existing Dial-Up Networking connections.


Connect serial cable to SPA and start Dial-up Networking.
When the connection is established, press any key to continue
configuring the array.
Connection to SPA (128.221.252.1) has been established and
verified.

Waiting to ensure SP is up and stable ...


NOTE: Some operations, like changing the serial number,
result in multiple reboots. This is expected.

Checking 128.221.252.1
SP at 128.221.252.1 has responded. Waiting 60 seconds to
ensure it stays up.
SP at 128.221.252.1 is up.

Configuring CX Series ...

Creating the hotspare ...

Setting Failover Mode

When prompted to create Celerra Control Volumes, type N:

Do you want to configure Celerra Control Volumes? [Y,N,] : N

Configuring cache

Configuring the cache ...

Cache Configured.

Tasks complete. The serial connection to the SP can now be


disconnected.

This procedure was logged into file:


C:\DOCUME~1\ADMINI~1\LOCALS~1\Temp\laptop
_script.log

When the script completes, type exit to close the Command Prompt window.

150 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Create data LUNs

Step Action

1. Log in to the Celerra Network Server.


Change to root (su) and type the password.

2. Determine the CXxxx serial number from a DOS window by typing:


$ navicli -h <SPA_IP_address> getagent
Output:
Agent Rev: 6.6.0 (3.1)
Name: K10
Desc:
Node: A-APM00035106458
Physical Node: K10
Signature: 979596
Peer Signature: 885857
Revision: 2.06.500.4.004
SCSI Id: 0
Model: 500
Model Type: Rackmount
Prom Rev: 3.00.00
SP Memory: 2048
Serial No: APM00035106458
SP Identifier: A
Cabinet: DPE2

Using Celerra Replicator (V1) Version 5.6.47 151 of 172


Step Action

3. Run the setup script using this command syntax:


$ /nas/sbin/setup_clariion <CXxxx_serial_number>
Use the serial number from the previous step.

Output:
# nas//setup_clariion APM00035106458

CLARIION(s) APM00035106458 will be setup.

Setup CLARiiON APM00035106458 storage device...

Enter the ip address for A_APM00035106458: 172.24.168.72

Enter the ip address for B_APM00035106458: 172.24.168.73


System 172.24.168.72 is up
System 172.24.168.73 is up

Clariion Array: APM00035106458 Model: CX500 Memory: 2048

Committed Base software package

The following 5 template(s) available:

1. CX_All_4Plus1_Raid_5
2. CX_Standard_Raid_5
3. CX_Standard_Raid_1
4. CX_Standard_Raid_5_Legacy
5. CX_Standard_Raid_1_Legacy
Please select a template in the range of 1-5 or 'q' to quit: 1
Summary:
2 disk group(s) are created. 8,9
5 spare(s) are created. 200,201,202,203,204
Enclosure(s) 0_0 are installed in the system.
Enclosure info:
----------------------------------------------------------------
0 1 2 3 4 5 6 7 8 9 10 11 12 13 14
----------------------------------------------------------------
0_0: 146 146 146 146 146 146 146 146 146 146 146 146 146 146 146
*8 *8 *8 *8 *8 *HS *9 *9 *9 *9 *9 *HS *HS *HS *HS
----------------------------------------------------------------
"*" indicates a diskgroup/spare which will be configured
Size Type Disks Spares
-------------------------
146 FC 15 5

Do you want to continue and configure as shown [yes or no]?: yes

Enclosure 0_0.
Created disk group 8, luns 16,17
Created spare 200
Created disk group 9, luns 18,19
Created spare 201
Created spare 202
Created spare 203
Created spare 204

Binding complete.
All luns are created successfully!

Enclosure(s) 0_0 are installed in the system.

152 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Step Action

3. Enclosure info:
----------------------------------------------------------------
0 1 2 3 4 5 6 7 8 9 10 11 12 13 14
----------------------------------------------------------------
0_0: 146 146 146 146 146 146 146 146 146 146 146 146 146 146 146
8 8 8 8 8 HS 9 9 9 9 9 HS HS HS HS
----------------------------------------------------------------

Configuration completed!

Setup of CLARiiON APM00035106458 storage device complete.

4. Register the World Wide Names using the Connectivity Status window in Navisphere.

5. Proceed with "Disk transport method" on page 140 procedure.

Using Celerra Replicator (V1) Version 5.6.47 153 of 172


Troubleshooting Celerra Replicator
As part of an effort to continuously improve and enhance the performance and
capabilities of its product lines, EMC periodically releases new versions of its
hardware and software. Therefore, some functions described in this document
might not be supported by all versions of the software or hardware currently in use.
For the most up-to-date information on product features, refer to your product
release notes.
If a product does not function properly or does not function as described in this
document, please contact your EMC Sales Representative.

Where to get help


Product information – For documentation, release notes, software updates, or for
information about EMC products, licensing, and service, go to the EMC Powerlink
website (registration required) at http://Powerlink.EMC.com.
Troubleshooting – For troubleshooting information, go to Powerlink, search for
Celerra Tools, and select Celerra Troubleshooting from the navigation panel on
the left.
Technical support – For technical support, go to Powerlink and choose Support.
On the Support page, you can access Support Forums, request a product
enhancement, talk directly to an EMC representative, or open a service request. To
open a service request, you must have a valid support agreement. Please contact
you EMC sales representative for details about obtaining a valid support agreement
or to answer any questions about your account.

Note: Do not request a specific support representative unless one has already been
assigned to your particular system problem.

Problem Resolution Roadmap for EMC Celerra contains additional information


about using Powerlink and resolving problems.

E-Lab Interoperability Navigator


The EMC E-Lab Interoperability Navigator is a searchable, web-based application
that provides access to EMC interoperability support matrices. It is available on the
EMC Powerlink website at http://Powerlink.EMC.com. After logging in to Powerlink,
go to Support > Interoperability and Product Lifecycle Information > E-Lab
Interoperability Navigator.

Log files for troubleshooting


The following log files are available to help troubleshoot replication:
◆ server_log for messages
◆ /nas/log/sys_log for messages
◆ /nas/log/cmd_log for internal commands run for replication
Additionally, use the fs_replicate –info –verbose <number_of_lines> command to
view the delta-set status.

154 of 172 Version 5.6.47 Using Celerra Replicator (V1)


server_log messages
Table 13 on page 155 shows an example message generated in the server_log for
a Data Mover performing replication at the source site:
Replication::Valid v:CISn363 Delta:2564 ad:525312 g:2564 nc:1
Table 13 Sample server_log message from source site

v: Volume ID of the source file system. In the above example, 363 is the volume
ID.

Delta: Delta-set number copied to the SavVol.

ad: Address at the block level (1 block = 512 bytes) on the SavVol where the delta
set is created.

g: Chunk number on the SavVol. If all delta sets consist of one chunk, this number
is the same as the delta-set number.

nc: Number of chunks in the delta set. One chunk equals 128 MB.

Table 14 on page 155 shows an example message generated in the server_log for
a Data Mover performing replication at the destination site:
Playback: v:361, Delta:3557, g:3557, ad:263168, nDelta:7
Table 14 Sample server_log message from destination site

v: Volume ID of the destination file system.

Delta: First delta-set number in the group replayed to the destination file system.

g: First chunk number in the first delta set. If all the delta sets consist of one
chunk, this number is the same as the delta-set number.

ad: Address on the SavVol where the first delta set in the group is located.

nDelta: Number of delta sets in the group replicated to the destination file system.

Using Celerra Replicator (V1) Version 5.6.47 155 of 172


Network performance troubleshooting
When you experience a performance issue during the transferring of delta sets,
check the following:
◆ Duplex network configuration mismatch (Full, Half, Auto, and so on) between
the Data Mover and the network switch
◆ Packet errors and input/output bytes using the server_netstat –i command on
the Data Mover
◆ Packet errors and input/output bytes on the network switch port
◆ Transfer rate in the fs_replicate -info command output
Create a file from the network client, refresh the source file system, and then check
the transfer rate in the fs_replicate -info command output.

Failure during transport of delta set


If a network failure occurs during the transport of a delta set over IP, the replication
service continues to resend this delta set until it is accepted by the destination site.

Failure of fs_copy command process


If the output of the fs_copy -list command reports failed in the Status field, or the
/nas/log/sys_log has one or more Copy failed messages, then the fs_copy process
has failed and you must do the following:
1. Abort the fs_copy operation by running the command fs_copy -abort.
2. Start replication from scratch, as described in "Start replication from scratch" on
page 129.
The following /nas/log/sys_log excerpt shows messages related to an fs_copy
failure:
Aug 29 23:00:38 2006 VMCAST:3:9 Slot 2: 1156910199:
Group:4673_0001854002020039
FSID:4666 Resync Copy failed. Full Copy has to be started. (t:17978065009336)
Aug 29 23:00:38 2006 VMCAST:3:2 Slot 2: 1156910199:
Group:4673_0001854002020039
FSID:4666 Copy Failed restartad:0x0 (t:17978065072171)

Control Station restarts during replication


During replication processing:
◆ If the Control Station restarts during an fs_copy, you must abort the file system
copy using the fs_copy -abort command for the source and destination file
systems. Then start the copy again.
◆ If the Control Station restarts with active file system replications, the replication
sessions are not affected and continue to run.

156 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Control Station fails over
For remote replication, if a primary Control Station fails over to a standby Control
Station, the replication service continues to run but the replication management
capabilities are unavailable. For example, you cannot perform list, abort, start, or
refresh replication functions from the standby Control Station. First resolve the
problem with the primary Control Station before executing any of these commands.
This applies to the Control Stations at the source and destination sites.

NS series loses power


If any of the NS series systems or a CNS-14 or NSX series system attached to a
CLARiiON storage system loses power, after you power up the system, determine if
you can restart replication using an incremental copy of the source file system as
described in "Restarting a replication relationship" on page 89. If not, abort as
detailed in "Abort Celerra Replicator" on page 79 and start replication from the
beginning as described in "Initiating replication" on page 39.

Return codes for fs_copy


This section describes the possible fs_copy return codes that can be helpful when
error checking scripts using the fs_copy command.
Table 15 on page 157 lists the possible return codes, and Table 16 on page 158
lists the return codes with corresponding error message IDs, a brief description,
and error severity. To obtain detailed information about a particular error, use the
nas_message -info <error_id> command.
Table 15 Return codes and description

Return code Description

0 Command completed successfully.

1 CLI usage error.

2 The object ID specified in the command line does not exist or


is invalid.

3 Unable to acquire locks or resources for the command to


start successfully.

4 Not applicable.

5 Possible communication error.

6 Transaction error reported during internal process


transaction or during object verification or post transaction
processing.

7 Data Mover error.

8 Not applicable.

Using Celerra Replicator (V1) Version 5.6.47 157 of 172


Table 16 fs_copy return codes with error IDs and message descriptions (page 1 of 3)

Return
Error ID Message description
code

0 10246 This warning indicates that the time skew between the local
and remote Celerra may have exceeded 10 minutes or that
there is a passphrase mismatch between the local and remote
Celerra.

0 10247 This warning indicates that the remote Celerra could not be
reached with an http connection.

1 2102 This CLI usage error indicates that an invalid command option,
such as the file system name, Celerra system name, interface
name, or IP address, was specified.

1 10142 This CLI usage error indicates that an invalid interface or IP


address was specified for the local or remote Celerra system.

2 2211 This error occurred because an invalid ID was specified.

3 2201 This error occurred because a process is not able to lock or


reserve a resource, such as a file system, Data Mover, or
interface.

5 4001 This is a communication error that occurred because the Data


Mover is not connected.

5 4002 This communication error occurred because either the source


Data Mover or the destination Data Mover (on which
destination file system is mounted) is not connected.

5 4036 This communication error indicates that the interface specified


for the source Data Mover is not connected to the interface on
the destination Data Mover.

6 2001 This error occurred during the transaction when another


process was modifying the database being accessed by the
current command.

6 2103 This CLI usage error occurred during argument verification and
indicates that an invalid command option, such as convert,
autofullcopy, qos, or resync, was specified.

6 2207 This error occurred when the transaction was aborted due to
an internal failure or abnormal termination of the command.

6 2225 This error indicates that an invalid checkpoint was specified in


the -fromfs option of the differential copy argument or that the
abort command was issued and the file system is involved in
multiple copy sessions. The copy session should be aborted
using the destination file system.

158 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Table 16 fs_copy return codes with error IDs and message descriptions (page 2 of 3)

Return
Error ID Message description
code

6 2227 This error indicates that the specified destination file system
has FLR status of Enterprise. FLR-C enabled file systems
(source or destination) are not supported for fs_copy or Celerra
Replicator (V1).

6 2237 This error occurred during the transaction as specified in the


error message.

6 2241 This error indicates that the command executed on the remote
system failed.

6 2243 This error indicates that the checkpoint specified in the


command is inactive.

6 2245 This error indicates that the file system type is invalid.

6 3105 This error indicates that either 1) an invalid file system is


specified for the copy session or the copy session already
exists, 2) the file system is not part of the specified session or,
3) the restart checkpoint specified is not newer than the
destination file system.

6 3128 This error indicates that the source file system was restored
from the checkpoint.

6 3134 This error indicates that the checkpoint specified is not newer
than the destination file system.

6 3136 This error occurred because the destination file system has
replication set up.

6 3138 This error indicates that the file system is already part of
another copy session.

6 3139 This error indicates that the file system is not part of a copy
session.

6 4019 This error occurred when polling the progress of the copy
session and indicates that the command failed to complete.

6 4103 This error occurred because the file system is not mounted.

6 4109 This error occurred because the file system is not mounted.

6 4111 This error occurred because the file system is mounted


read/write by the server specified in the message.

6 4124 This error occurred because the file system is mounted


read/write by the server specified in the error.

Using Celerra Replicator (V1) Version 5.6.47 159 of 172


Table 16 fs_copy return codes with error IDs and message descriptions (page 3 of 3)

Return
Error ID Message description
code

6 4205 This error occurred because there is an invalid interface


specified for the source file system. Check the interfaces
specified for the Data Mover on which the file system is
mounted.

6 4420 This error occurred because the destination file system is a


backup of another replication or copy session.

6 4424 This error occurred because fsck is being executed on the file
system.

6 4425 This error occurred because aclck is being executed on the file
system.

6 4446 This error occurred because the remote Celerra system is


running an out-of-family NAS code version with respect to the
NAS code version running on the local Celerra system.

6 4447 This error occurred because IP Alias is configured on the


system and the remote system's version could not be
identified.

6 5000 This error occurred during the transaction because of an


internal error.

6 10233 This error indicates that the query executed on the remote
system failed.

6 10272 This error occurred because the destination file system could
not found.

6 10273 This error indicates that the copy session failed.

6 10274 This error occurred because either the user aborted the copy
session or there is a problem with the Data Mover.

6 10277 This error indicates that the checkpoint specified is older than
than the replication configured on the source file system.

6 10299 This error occurred because a writable checkpoint is specified


in the CLI.

6 10311 This error indicates that the clean up process for the copy
session on destination failed.

6 10312 This error occurred because the size of the source and
destination file systems does not match.

7 10310 This error occurred because the system was unable to retrieve
copy session information from the Data Mover.

160 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Error messages for Celerra Replicator
As of version 5.6, all new event, alert, and status messages provide detailed
information and recommended actions to help you troubleshoot the situation.
To view message details, use any of the following methods:
◆ Celerra Manager:
• Right-click an event, alert, or status message and select to view Event
Details, Alert Details, or Status Details.
◆ Celerra CLI:
• Use the nas_message -info <error_id> command to retrieve the detailed
information for a particular error.
◆ EMC Celerra Network Server Error Messages Guide:
• Use this guide to locate information about messages that are in the earlier-
release message format.
◆ Powerlink:
• Use the text from the error message’s brief description or the message’s ID
to search the Knowledgebase on Powerlink. After logging in to Powerlink, go
to Support > Knowledgebase Search > Support Solutions Search.

Using Celerra Replicator (V1) Version 5.6.47 161 of 172


Related information
Specific information related to the features and functionality described in this
document is included in:
◆ EMC Celerra Glossary
◆ EMC Celerra Network Server Command Reference Manual
◆ EMC Celerra Network Server Error Messages Guide
◆ EMC Celerra Network Server Parameters Guide
◆ EMC Celerra Network Server Release Notes
◆ Configuring EMC Celerra Events and Notifications
◆ Configuring EMC Celerra Time Services
◆ Managing EMC Celerra Volumes and File Systems Manually
◆ Managing EMC Celerra Volumes and File Systems with Automatic Volume
Management
◆ Online Celerra man pages
◆ Problem Resolution Roadmap for EMC Celerra
◆ Replicating EMC Celerra CIFS Environments (V1)
◆ Using EMC Celerra FileMover
◆ Using EMC Celerra Replicator for iSCSI (V1)
◆ Using File-Level Retention on EMC Celerra
◆ Using International Character Sets with EMC Celerra
◆ Using SnapSure on EMC Celerra
◆ Using TimeFinder/FS, NearCopy, and FarCopy with EMC Celerra
The EMC Celerra Network Server Documentation CD, supplied with Celerra and
also available on the EMC Powerlink website, provides the complete set of EMC
Celerra customer publications. After logging in to Powerlink, go to Support >
Technical Documentation and Advisories > Hardware/Platforms
Documentation > Celerra Network Server. On this page, click Add to Favorites.
The Favorites section on your Powerlink home page provides a link that takes you
directly to this page.
Celerra Support Demos are available on Powerlink. Use these instructional videos
to learn how to perform a variety of Celerra configuration and management tasks.
After logging in to Powerlink, go to Support > Product and Diagnostic Tools >
Celerra Tools > Celerra Support Demos.

162 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Training and Professional Services
EMC Customer Education courses help you learn how EMC storage products work
together within your environment in order to maximize your entire infrastructure
investment. EMC Customer Education features online and hands-on training in
state-of-the-art labs conveniently located throughout the world. EMC customer
training courses are developed and delivered by EMC experts. Go to EMC
Powerlink at http://Powerlink.EMC.com for course and registration information.
EMC Professional Services can help you implement your Celerra Network Server
efficiently. Consultants evaluate your business, IT processes, and technology and
recommend ways you can leverage your information for the most benefit. From
business plan to implementation, you get the experience and expertise you need,
without straining your IT staff or hiring and training new personnel. Contact your
EMC representative for more information.

Using Celerra Replicator (V1) Version 5.6.47 163 of 172


Appendix A: fs_replicate -info output fields
The first part of the output from the fs_replicate -info command presents information
about the source file system. Table 17 on page 164 describes these output fields.
Table 17 First section of fs_replicate -info output fields

Field Description

id Source file system ID.

name Source file system name.

fs_state Condition of the source file system. Indicates whether the


source file system is active (active), mounted read-only
(romounted), or unmounted (frozen).

type Whether it is a replication or playback service.

replicator_state Indicates if the replication service is active, inactive, or


creating a delta set. If inactive, replication has fallen out-of-
sync.

source_policy How replication handles flow-control situations. Values are


ReadOnly, Freeze, or NoPolicy, where:
ReadOnly = the file system is only available for reads.
Freeze = reads and writes are not allowed.
NoPolicy = no policy is defined.

high_water_mark Size in MB of the file system changes accumulated since the


last delta set was created. When this size is reached, the
replication service automatically creates a delta set on the
SavVol.

time_out Interval (in seconds) when the replication service creates a


delta set.

current_delta_set Current delta set being processed. This reflects the current
delta set being tracked in memory.

current_number_of_block Number of modified blocks in the current delta set. One block
s size is 8 KB.

flow_control Indicates if there is enough space in the SavVol to write a new


delta set to the SavVol.

Note: An active status indicates there is not enough SavVol


space to process the next delta set.

total_savevol_space The total size of the SavVol.

savevol_space_available The amount of free space in the SavVol.

164 of 172 Version 5.6.47 Using Celerra Replicator (V1)


The second part of the output from the fs_replicate command presents information
about the destination file system. Table 18 on page 165 describes these output
fields.
Table 18 Second section of fs_replicate -info output fields

Field Description

id Destination file system id.

name Destination file system name.

type Indicates whether it is a replication or playback service.

playback_state Indicates whether the playback service is active, inactive, or


replaying a delta set. If inactive, replication has not started.

high_water_mark Size in MB of the delta sets at which the replication service


automatically replays a delta set on the SavVol.

time_out Interval (in seconds) when the replication service creates or


replays a delta set.

current_delta_set Lists the next delta set to replay to the destination file system.

flow_control Indicates whether there is enough space to write an incoming


delta set to the destination SavVol.

Note: An active status indicates there is not enough SavVol


space to process the next delta set.

total_savevol_space The total size of the SavVol.

savevol_space_available The amount of free space in the SavVol.

outstanding delta sets The delta sets on the destination SavVol the replication
service has not played back to the destination file system.

Using Celerra Replicator (V1) Version 5.6.47 165 of 172


The third part of the output from the fs_replicate command presents information on
the communication state between the source and destination Data Movers.
Table 19 on page 166 describes these output fields.
Table 19 Third section of fs_replicate -info output fields

Field Description

communication_state Indicates whether the Data Movers on each Celerra system


can communicate with each other.

current_transfer_rate The rate at which the last 128 MB of data was sent across the
IP network.

avg_transfer_rate The average rate at which the last 128 sets of data were sent
across the IP network.

source_ip IP address of the source Data Mover.

source_port Port number of the source Data Mover. This is assigned


dynamically.

destination_ip IP address of the destination Data Mover.

destination_port Port number of the destination Data Mover. This is always


assigned the number 8888 by default.

QOS_bandwidth The IP network bandwidth throttle used for this replication


relationship.
Zero (0) represents maximum available network bandwidth.

When using the verbose option with the fs_replicate -info command, the replication
service generates this additional output. Table 20 describes these output fields.
Table 20 fs_replicate -info -verbose output fields (page 1 of 2)

Field Description

Source file system

Delta Delta-set number or delta-set ID that was created.

Create Time Date and start time the delta set was created.

Dur Duration of time (in seconds) to create the delta set.

Blocks Number of blocks modified in the delta set.

166 of 172 Version 5.6.47 Using Celerra Replicator (V1)


Table 20 fs_replicate -info -verbose output fields (page 2 of 2)

Field Description

Destination file system

Playback Time Date and start time the delta set replayed to the destination file
system.

Dur Duration of time (in seconds) to replay the delta set or DsinGroup.

Blocks Number of blocks modified in the delta set.

DSinGroup Number of delta sets in the group played back. In some instances,
the playback service can play back more than one delta set at
once. In this case, the Dur and Blocks fields refer to the group as a
whole, not an individual delta set.

Using Celerra Replicator (V1) Version 5.6.47 167 of 172


168 of 172 Version 5.6.47 Using Celerra Replicator (V1)
nas_cel, verify Control Station relationship 18, 41
Index nas_cel, view passphrase 42
nas_fs, creating destination 45
server_mount, mounting file systems 45
A server_mountpoint, creating 45
aborting replication 79 server_sysstat, monitor memory usage 35
Automatic File System Extension 98, 100, 104 configuration
definition 5 considerations 35
extending after replication failure 104 replication policies 30
recovering from failure 100 Control Station
starting replication from scratch 104 preinitialization 40

B D
bandwidth size data flow control 31
changing policy 109 delta set
modifying 109 checking status 115
definition 6
minimum size 30
C overview 30
cautions delta-set
graceful shutdown 9 transport failure 156
serial replication sessions 9 disk transport 140
system 9
Unicode to ASCII replication 9
Celerra Replicator E
cautions 9 exit codes for fs_copy 157
checking status of 55, 114 extending file system
log files 154 after replication failover 104
restarting replication 89
restrictions 7 F
starting replication 48 failover
system requirements 22 initiate 64
upgrade considerations 24, 25 options 62
checkpoint failure
definition 6 delta-set transport 156
Commands failure, fs_replicate 50
fs_replicate, flow control options 32 file system
nas_fs, calculate SavVol size 33 automatically extending size 98
commands events 121
fs_ckpt, copy checkpoints 50 manually extending size 101
fs_ckpt, creating 43 flow control
fs_ckpt, using 59 freeze 107
fs_copy, checkpoint copy 46 read-only 107
fs_copy, copy checkpoints 45 fs_copy return codes 157
fs_copy, copy incremental changes 52
fs_copy, events for 122
fs_copy, using 46, 52 H
fs_replicate, aborting 79 high water mark
fs_replicate, changing bandwidth 109 definition 6
fs_replicate, check status 55 resetting policy 105
fs_replicate, control SavVol size 34 HTTP communication 17
fs_replicate, description of output fields 165
fs_replicate, failover 64 I
fs_replicate, failure 50
information, related 163
fs_replicate, output definitions 165
Initial Copy
fs_replicate, resynchronize 68
transporting by disk 144
fs_replicate, reverse 76
initial copy
fs_replicate, starting 48
transporting by disk 140
fs_replicate, suspending 81
transporting by tape 144
fs_replicate, using 57

Using Celerra Replicator (V1) Version 5.6.47 169 of 172


IP replication service description of 11
definition 6 establish communication between Celerra systems 17
process overview 12
start replication 48
L
system requirements 22
local replication
verify communication between Celerra systems 41
definition 6
replication
description of 10
aborting 79
process overview 10
starting from scratch 104
system requirements 22
replication failover
log files 154
definition 6
loopback replication, definition 6
process overview 13
losing power, CLARiiON storage system 157
replication policy
configuring 35
M flow control 107
messages, server_log 155 high watermark 30
setting 105
time-out 30
N triggers 35
nas_cel, using 41 replication service, definition 6
nas_fs, using 45, 101 replications per Data Mover 34
network performance, troubleshooting 156 Replicator SavVol
NS series, loses power 157 definition 6
parameter for changing size 123
O size requirements 33
overview restart
local replication 10 out of synchronization
remote replication 12 checkpoints 21, 59
replication failover 13 suspended replication 20
replication reverse 17 restrictions
suspend 81 system 7
TimeFinder/FS 7
TimeFinder/FS Near and Far Copy 7
P resynchronize
parameters, changing Replicator SavVol size 123 process overview 16
passphrase reversing replication
establishing 17 after failover 75
viewing 42 definition 6
physical transport of replication data 139 maintenance procedure 111
playback service process overview 17
checking 115
definition 6
policy S
bandwidth size 109 server_df, monitoring file systems 114
flow control 31 server_log 155
high watermark 30 server_param command, SavVol default 123
time-out 30 server_sysstat, monitoring replication 114
SnapSure, SavVol definition 6
SNMP traps 121
R status
recovering from a corrupted file system using nas_fsck 130 checking 55
recovering from auto extension failure 100 of replication 114
remote replication suspend
copy overview 81
checkpoint 45 procedure 20
incremental changes 52 system
create requirements 22
checkpoint 42 restrictions 7
destination file system 45
definition 6

170 of 172 Version 5.6.47 Using Celerra Replicator (V1)


T
tape, transporting initial copy procedure 144
timeout
definition 6
replication policy 30
resetting policy 105
triggers, data flow 31
troubleshooting
accommodating network concerns 127
calculating modifications rate on source file system 127
changing the passphrase 124
Control Station failovers 157
Control Station reboots during replication 156
copying a file system to multiple destinations with fs_copy 136
creating checkpoints on the destination site 136
creating restartable checkpoints 125
enlarging SavVol size 126
error messages 162
failure during transport of delta set 156
network performance 156
NS series loses power 157
reading server_log messages 155
recovering from an inactive replication state 135
return codes for fs_copy 157
starting replication from a differential copy 128
starting replication from scratch 129
temporarily mount destination file system read/write 133
using log files 154
trust relationship 17

U
upgrading 24, 25

V
Virtual Data Mover, definition 7
virtual provisioning
definition 7

Using Celerra Replicator (V1) Version 5.6.47 171 of 172


About this document
As part of its effort to continuously improve and enhance the performance and capabilities of the Celerra Network Server product line, EMC
periodically releases new versions of Celerra hardware and software. Therefore, some functions described in this document may not be
supported by all versions of Celerra software or hardware presently in use. For the most up-to-date information on product features, see your
product release notes. If your Celerra system does not offer a function described in this document, contact your EMC Customer Support
Representative for a hardware upgrade or software update.
Comments and suggestions about documentation
Your suggestions will help us improve the accuracy, organization, and overall quality of the user documentation. Send a message to
techpubcomments@EMC.com with your opinions of this document.

Copyright © 1998-2009 EMC Corporation. All rights reserved.


EMC believes the information in this publication is accurate as of its publication date. The information is subject to change without notice.
THE INFORMATION IN THIS PUBLICATION IS PROVIDED “AS IS.” EMC CORPORATION MAKES NO REPRESENTATIONS OR
WARRANTIES OF ANY KIND WITH RESPECT TO THE INFORMATION IN THIS PUBLICATION, AND SPECIFICALLY DISCLAIMS
IMPLIED WARRANTIES OF MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE.
Use, copying, and distribution of any EMC software described in this publication requires an applicable software license.
For the most up-to-date regulatory document for your product line, go to the Technical Documentation and Advisories section on EMC
Powerlink.
For the most up-to-date listing of EMC product names, see EMC Corporation Trademarks on EMC.com.
All other trademarks used herein are the property of their respective owners.

172 of 172 Version 5.6.47

Vous aimerez peut-être aussi