Top Banner
1 of 172 Contents Introduction to Celerra Replicator . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5 Terminology . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5 Restrictions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7 Cautions. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9 Celerra Replicator concepts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10 Local replication . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10 Remote replication . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11 Activating the destination file system as read/write . . . . . . . . . . . . . 13 Communication between Celerra Network Servers. . . . . . . . . . . . . . 17 How resynchronization works . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18 How suspend works . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19 How replication relationship restarts . . . . . . . . . . . . . . . . . . . . . . . . . 20 System requirements for Celerra Replicator . . . . . . . . . . . . . . . . . . . . . . . 22 Local replication . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22 Remote replication . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22 Upgrading from previous Celerra Network Server versions . . . . . . . . . . 24 Upgrade from a version earlier than 5.5.39.2 . . . . . . . . . . . . . . . . . . . 24 Upgrade from Celerra Network Server version 5.5.39.2 or later. . . . 25 Planning considerations for Celerra Replicator . . . . . . . . . . . . . . . . . . . . 29 Replication policies . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29 SavVol size requirements for remote replication . . . . . . . . . . . . . . . 33 Determine the number of replications per Data Mover . . . . . . . . . . . 34 Configuration considerations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35 User interface choices for Celerra Replicator . . . . . . . . . . . . . . . . . . . . . . 37 Roadmap for Celerra Replicator . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38 Initiating replication . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 39 Task 1: Establish communication . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40 Task 2: Verify communication . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40 Task 3: Create SnapSure checkpoint of source file system. . . . . . . 42 Task 4: Create the destination file system . . . . . . . . . . . . . . . . . . . . . 45 Task 5: Copy checkpoint to the destination file system . . . . . . . . . . 45 Task 6: Begin replication . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48 Task 7: Create a second checkpoint of the source file system . . . . 50 Task 8: Copy incremental changes . . . . . . . . . . . . . . . . . . . . . . . . . . . 52 Using Celerra Replicator (V1) P/N 300-004-184 Rev A08 Version 5.6.47 December 2009
172

Celerra_Replcatr

Nov 22, 2014

Download

Documents

rogerbwilson
Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: Celerra_Replcatr

1 of 172

Contents

Introduction to Celerra Replicator . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .5Terminology . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .5Restrictions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .7Cautions. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .9

Celerra Replicator concepts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .10Local replication . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .10Remote replication . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .11Activating the destination file system as read/write . . . . . . . . . . . . .13Communication between Celerra Network Servers. . . . . . . . . . . . . .17How resynchronization works . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .18How suspend works . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .19How replication relationship restarts . . . . . . . . . . . . . . . . . . . . . . . . .20

System requirements for Celerra Replicator. . . . . . . . . . . . . . . . . . . . . . .22Local replication . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .22Remote replication . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .22

Upgrading from previous Celerra Network Server versions . . . . . . . . . .24Upgrade from a version earlier than 5.5.39.2 . . . . . . . . . . . . . . . . . . .24Upgrade from Celerra Network Server version 5.5.39.2 or later. . . .25

Planning considerations for Celerra Replicator . . . . . . . . . . . . . . . . . . . .29Replication policies . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .29SavVol size requirements for remote replication . . . . . . . . . . . . . . .33Determine the number of replications per Data Mover . . . . . . . . . . .34Configuration considerations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .35

User interface choices for Celerra Replicator. . . . . . . . . . . . . . . . . . . . . .37Roadmap for Celerra Replicator . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .38Initiating replication. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .39

Task 1: Establish communication . . . . . . . . . . . . . . . . . . . . . . . . . . . .40Task 2: Verify communication . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .40Task 3: Create SnapSure checkpoint of source file system. . . . . . .42Task 4: Create the destination file system . . . . . . . . . . . . . . . . . . . . .45Task 5: Copy checkpoint to the destination file system. . . . . . . . . .45Task 6: Begin replication . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .48Task 7: Create a second checkpoint of the source file system . . . .50Task 8: Copy incremental changes. . . . . . . . . . . . . . . . . . . . . . . . . . .52

Using Celerra Replicator (V1)P/N 300-004-184

Rev A08

Version 5.6.47

December 2009

Page 2: Celerra_Replcatr

Using Celerra Replicator (V1)Version 5.6.47 2 of 172 Using Celerra Replicator (V1)Version 5.6.47 2 of 172

Task 9: Verify file system conversion . . . . . . . . . . . . . . . . . . . . . . . . 54Task 10: Check replication status . . . . . . . . . . . . . . . . . . . . . . . . . . . 55Task 11: Create restartable checkpoints . . . . . . . . . . . . . . . . . . . . . . 59

Recovering replication data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 61Task 1: Replication failover. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 62Task 2: Resynchronize the source and destination sites . . . . . . . . 66Task 3: Replication reversal . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 74

Abort Celerra Replicator. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 79Suspend a replication relationship . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 81

Verify the suspended replication relationship . . . . . . . . . . . . . . . . . 88Restarting a replication relationship . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 89

Verify that the replication relationship is not synchronized . . . . . . 90Restart replication relationship . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 91

Extending the size of a file system . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 98Extend file system size automatically . . . . . . . . . . . . . . . . . . . . . . . . 98Extend file system size . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 101

Resetting replication policy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 105High water mark and time-out policies . . . . . . . . . . . . . . . . . . . . . . 105Modify replication policy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 106Change flow-control policies . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 107Set bandwidth size . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 109Set policies using parameters . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 110

Reverse the direction of a replication relationship . . . . . . . . . . . . . . . . 111Verify the reverse direction of replication relationship . . . . . . . . . 113

Monitor replication . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 114Checking playback service and outstanding delta sets . . . . . . . . . . . . 115

Task 1: Determine playback service status . . . . . . . . . . . . . . . . . . . 115Task 2: Playback delta set. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 118Task 3: Verify delta set . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 119

Events for Celerra Replicator. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 121Change the Celerra Replicator SavVol default size . . . . . . . . . . . . . . . . 123Change the passphrase between Celerra Network Servers . . . . . . . . . 124Managing and avoiding IP replication problems . . . . . . . . . . . . . . . . . . 125

Preventive measures to avoid IP replication problems . . . . . . . . . 125Replication restart methods . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 128Recovering from a corrupted file system . . . . . . . . . . . . . . . . . . . . 130Managing anticipated destination site or network outages . . . . . . 131Managing unanticipated destination site or network outages. . . . 132Managing unanticipated source site outages . . . . . . . . . . . . . . . . . 133Managing expected source site outages . . . . . . . . . . . . . . . . . . . . . 133Mount the destination file system read/write temporarily . . . . . . . 133Recovering from an inactive replication state. . . . . . . . . . . . . . . . . 135Creating checkpoints on the destination site . . . . . . . . . . . . . . . . . 136Copy file system to multiple destinations with fs_copy. . . . . . . . . 136

Transporting replication data using disk or tape. . . . . . . . . . . . . . . . . . 139Disk transport method . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 140Tape transport method . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 144

Setting up the CLARiiON disk array . . . . . . . . . . . . . . . . . . . . . . . . . . . . 147Review the prerequisites . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 147Run the setup script . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 149Create data LUNs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 151

Troubleshooting Celerra Replicator . . . . . . . . . . . . . . . . . . . . . . . . . . . . 154Where to get help . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 154

Page 3: Celerra_Replcatr

3 of 172Version 5.6.47Using Celerra Replicator (V1)

E-Lab Interoperability Navigator . . . . . . . . . . . . . . . . . . . . . . . . . . . .154Log files for troubleshooting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .154server_log messages . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .155Network performance troubleshooting. . . . . . . . . . . . . . . . . . . . . . .156Failure during transport of delta set . . . . . . . . . . . . . . . . . . . . . . . . .156Failure of fs_copy command process . . . . . . . . . . . . . . . . . . . . . . .156Control Station restarts during replication . . . . . . . . . . . . . . . . . . .156Control Station fails over . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .157NS series loses power . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .157Return codes for fs_copy. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .157

Error messages for Celerra Replicator . . . . . . . . . . . . . . . . . . . . . . . . . .161Related information . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .162

Training and Professional Services . . . . . . . . . . . . . . . . . . . . . . . . .163Appendix A: fs_replicate -info output fields . . . . . . . . . . . . . . . . . . . . . .164Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .169

Page 4: Celerra_Replcatr

Using Celerra Replicator (V1)4 of 172 Version 5.6.47

Page 5: Celerra_Replcatr

5 of 172Version 5.6.47Using Celerra Replicator (V1)

Introduction to Celerra Replicator

EMC® Celerra Replicator™ produces a read-only, point-in-time copy of a source file system and periodically updates this copy, making it consistent with the source file system. This read-only copy can be used by a Data Mover in the same Celerra cabinet, or a Data Mover at a remote site for content distribution, backup, and application testing.

This technical module is part of the EMC Celerra® Network Server documentation set and is intended for system administrators who are responsible for IP replication in the Celerra environment. Before using Celerra Replicator, system administrators establishing replication should understand Celerra volumes and file systems.

This technical module is one of several technical modules that describe replication using different implementations of Celerra Replicator. Use the following guidelines to navigate the technical modules:

◆ Read this technical module to learn the basics about the Celerra Replicator product and how it performs local replication or remote replication of Production File Systems (PFSs) over an IP network to a destination. This technical module describes how to set up the IP environment for replication and how to use Celerra Replicator to replicate your file systems. Because Celerra Replicator relies on EMC SnapSure™ checkpoints (read-only, logical, point-in-time images) for the initial copy and for replication restart, you might find it useful to read Using SnapSure on EMC Celerra to learn more about checkpoints.

◆ Read Replicating EMC Celerra CIFS Environments (V1) to learn how to use Celerra Replicator to perform replication in the CIFS environment. Replicating EMC Celerra CIFS Environments (V1) describes how to replicate the CIFS environment information in the root file system of a Virtual Data Mover (VDM) as well as how to replicate the file systems mounted to that VDM. To learn about VDMs in general, read Configuring Virtual Data Movers for EMC Celerra.

◆ Read Using EMC Celerra Replicator for iSCSI (V1) to learn about iSCSI replication using Celerra Replicator for iSCSI product. Using EMC Celerra Replicator for iSCSI (V1) describes how to replicate production iSCSI LUNs by asynchronously distributing local point-in-time copies of the LUNs to a destination.

◆ Read Using EMC Celerra Replicator (V2) to learn about the new Celerra Replicator product.

◆ Read Managing EMC Celerra Volumes and File Systems with Automatic Volume Management and Managing EMC Celerra Volumes and File Systems Manually to learn more about establishing replication.

Terminology

This section defines terms important to understanding replication on the Celerra Network Server. The EMC Celerra Glossary provides a complete list of Celerra terms.

Automatic File System Extension: A configurable Celerra file system feature that automatically extends a file system created or extended with Automatic Volume

Page 6: Celerra_Replcatr

Using Celerra Replicator (V1)6 of 172 Version 5.6.47

Manager (AVM) when the high water mark (HWM) is reached. See also high water mark.

checkpoint: Read-only, logical, point-in-time image of a file system. A checkpoint is sometimes referred to as a checkpoint file system or a SnapSure file system.

delta set: Set containing the block modifications made to the source file system that Celerra Replicator uses to update the destination file system (read-only, point-in-time, consistent replica of the source file system). The minimum delta-set size is 128 MB.

high water mark (HWM): The trigger point at which Celerra Network Server performs one or more actions, such as sending a warning message, extending a volume, or updating a replicated file system, as directed by the related feature’s software/parameter settings.

IP replication service: The service that uses the IP network to transfer the delta sets from the replication SavVol on the source site to the replication SavVol on the destination site.

local replication: Replication of a file system on a single Celerra Network Server with the source file system on one Data Mover and the destination file system on another Data Mover.

loopback replication: Replication of a file system with the source and destination file systems residing on the same Data Mover.

playback service: The process of reading the delta sets from the destination SavVol and updating the destination file system.

remote replication: Replication of a file system from one Celerra Network Server to another. The source file system resides on a different Celerra system from the destination file system.

replication: A service that produces a read-only, point-in-time copy of a source file system. The service periodically updates the copy, making it consistent with the source file system.

replication reversal: The process of reversing the direction of replication. The source file system becomes read-only and the destination file system becomes read/write.

replication service: Service that copies modified blocks from the source file system to a replication SavVol prior to transferring the data to the destination file system.

Replicator ConfigVol: An internal information store for replication. Provides a storage vehicle for tracking changes in the source file system.

Replicator failover: The process that changes the destination file system from read-only to read/write and stops the transmission of replicated data. The source file system, if available, becomes read-only.

Replicator SavVol: A Celerra volume, required by replication, used to store modified data blocks from the source file system.

SnapSure SavVol: A Celerra volume to which SnapSure copies point-in-time data blocks from the PFS before the blocks are altered by a transaction. SnapSure uses the contents of the SavVol and the unchanged PFS blocks to maintain a checkpoint of the PFS.

timeout: Time interval at which the system takes a predetermined action.

Page 7: Celerra_Replcatr

7 of 172Version 5.6.47Using Celerra Replicator (V1)

Virtual Data Mover (VDM): A Celerra software feature enabling users to administratively separate CIFS servers, replicate CIFS environments, and move CIFS servers from Data Mover to Data Mover.

Virtual Provisioning: A configurable Celerra file system feature that can only be used in conjunction with Automatic File System Extension. This option lets you allocate storage based on longer-term projections, while you dedicate only the file system resources you currently need. Users—NFS or CIFS clients and application —see the virtual maximum size of the file system of which only a portion is physically allocated. Combined, the Automatic File System Extension and Virtual Provisioning options let you grow the file system gradually on an as-needed basis.

Restrictions

The following restrictions apply to Celerra Replicator:

◆ Celerra Data Migration Service (CDMS) is unsupported (an mgfs file system cannot be replicated).

◆ Multi-Path File System (MPFS) is supported on the source file system, but not on the destination file system.

◆ EMC E-Lab™ Interoperability Navigator provides information about disaster recovery replication products such as EMC SRDF®/Synchronous (SRDF/S) and SRDF/Asynchronous (SRDF/A).

◆ For EMC TimeFinder®/FS:

• A business continuance volume (BCV) cannot be a source or a destination file system for replication. You can replicate the underlying source file system, but not the BCV.

• Do not use the TimeFinder/FS -Restore option for a replicated source file system. Replication will be unaware of any changes because these changes occur at the volume level. However, you can restore on a single file basis using a NFS/CIFS client, which has access to the source file system and the BCV of the source file system.

• Do not use TimeFinder/FS with a file system that was created on a slice volume. Creating a file system using the samesize option slices the volume. TimeFinder does not recognize sliced partitions. Using TimeFinder/FS, NearCopy, and FarCopy with EMC Celerra further details this feature.

◆ For TimeFinder/FS Near Copy and Far Copy:

• A BCV cannot be a source or a destination file system for replication. You can replicate the underlying source file system, but cannot replicate the BCV.

◆ Do not extend the source file system while fs_copy is running.

◆ On a per-Data-Mover basis, the total size of all file systems, the size of all SavVols used by SnapSure, and the size of all SavVols used by the Celerra Replicator feature must be less than the total supported capacity of the Data Mover. The EMC Celerra Network Server Release Notes, available at http://Powerlink.EMC.com, the EMC Powerlink® website, provide a list of Data Mover capacities.

Page 8: Celerra_Replcatr

Using Celerra Replicator (V1)8 of 172 Version 5.6.47

◆ When replicating databases, additional application-specific actions may be necessary to bring the database to a consistent state (for example, quiescing the database).

◆ If you plan to enable international character sets (Unicode) on your source and destination sites, you must first set up translation files on both sites before starting Unicode conversion on the source site. Using International Character Sets with EMC Celerra describes this action in detail.

◆ In the case of multiple file systems, all fs_replicate commands must be executed sequentially.

◆ IP replication failover is not supported for local groups unless you use VDMs. Configuring Virtual Data Movers for EMC Celerra describes VDM configuration.

◆ IP replications created prior to Celerra Network Server version 5.5 in which the source file system contained iSCSI LUNs are no longer supported. These replications will continue to run in their current state, but you cannot actively manage (suspend or resume) them. Any attempt to perform such an operation prompts an error stating that the “item is currently in use by iSCSI.” You can abort or delete the replication. EMC recommends that you convert any existing IP replications to iSCSI replications as soon as possible.

◆ Do not use IP aliasing for IP replication. Use IP aliasing only for Control Station access from the client.

◆ For full management capability, ensure that you have the same Celerra version on the source and destination Celerra Network Servers. For example, you cannot have version 5.5 on the source side and version 5.6 on the destination side.

For limited management capability and no fs_copy support, out-of-family replication support is available and requires NAS version 5.5.39.2 or later on the source Celerra and 5.6.47 or later on the destination Celerra. "Upgrade from Celerra Network Server version 5.5.39.2 or later" on page 25 provides more details.

◆ For FLR-E-enabled file systems, both the source and destination file systems must have the same FLR setting enabled. For example, if the source file system has FLR-E enabled, then the destination file system must also have FLR-E enabled. If the source file system does not have FLR-E enabled, then the destination cannot have FLR-E enabled.

◆ File systems enabled for processing by Celerra Data Deduplication cannot be replicated using Celerra Replicator. In addition, you cannot enable deduplication on a file system that is already being replicated by Celerra Replicator.

◆ In Celerra Replicator (V1) version 5.6, the fs_replicate -refresh command does not create a new delta set until the production file system is updated. Consequently, policy values for timeout, high water mark (HWM), autofreeze, autoreadonly, and so on, that are modified using the fs_replicate -modify or fs_replicate -refresh command, remain ineffective until a new delta set is created.

In versions prior to 5.6, the fs_replicate -refresh command creates a new delta set even if no updates are made to the PFS.

Page 9: Celerra_Replcatr

9 of 172Version 5.6.47Using Celerra Replicator (V1)

Cautions

This section lists the cautions for using this feature on Celerra Network Server. If any of this data is unclear, contact EMC Customer Service for assistance:

◆ To provide a graceful shutdown in an electrical power loss, Celerra Network Server and the storage array need to have Uninterruptible Power Supply (UPS) protection. If this is not provided, replication will become inactive and might result in data loss. If replication becomes inactive, consult "Restarting a replication relationship" on page 89 to determine if you can resume the replication relationship.

◆ Replicating file systems from a Unicode-enabled Data Mover to an ASCII-enabled Data Mover is not supported. I18N mode (Unicode or ASCII) must be the same on the source and destination Data Movers.

◆ Replication sessions should run serially, not concurrently. That is, they should start one after the other, not simultaneously.

Page 10: Celerra_Replcatr

Using Celerra Replicator (V1)10 of 172 Version 5.6.47

Celerra Replicator concepts

This section explains how replication produces and uses a read-only, point-in-time copy of a source file system on the same or different Celerra Network Servers. It also describes how Celerra Replicator enables you to activate failover to the destination site for production, if the source site experiences a disaster and is unavailable for data processing.

Local replication

Replication produces a read-only copy of the source file system for use by a Data Mover in the same Celerra cabinet. The source and destination file systems are stored on separate volumes. Local replication can use different Data Movers or the same Data Mover.

Local replication process

Figure 1 on page 10 and subsequent steps show the processing that occurs when using local replication for the first time.

Figure 1 Local replication

CNS-000765

Sourcefile

system

Celerra Network Server

PrimaryData Mover

SavVol

Storage unit

Destinationfile

system

SecondaryData Mover

2

3 4

5

1

Page 11: Celerra_Replcatr

11 of 172Version 5.6.47Using Celerra Replicator (V1)

The process is as follows:

1. Throughout this process, network clients read and write to the source file systems through the primary Data Mover without interruption.

2. For the initial replication start, the source and destination file systems are manually synchronized using the fs_copy command.

3. After synchronization, the replication service uses the addresses of all block modifications made to the source file system to create one or more delta sets. The modified blocks are copied to the SavVol shared by the primary and secondary Data Movers.

4. The local replication playback service periodically reads any available, complete delta sets and updates the destination file system, making it consistent with the source file system. During this time, the system tracks all subsequent changes made to the source file system.

5. The secondary Data Mover exports the read-only copy to use for content distribution, backup, and application testing. This optional step is done manually.

Remote replication

Remote replication creates and periodically updates a read-only copy of a source file system at a remote (destination) site. This is done by transferring changes made to a source file system at a local site to a file system replica (destination) at the destination site over an IP network. These transfers are automatic and are based on user-definable replication policies.

Page 12: Celerra_Replcatr

Using Celerra Replicator (V1)12 of 172 Version 5.6.47

Remote replication process

Figure 2 on page 12 and subsequent steps show the processing that occurs when using remote replication for the first time:

Figure 2 Remote replication

1. Throughout this process, network clients read and write to the source file systems through a Data Mover at the source site without interruption.

2. For the initial replication start, the source and destination file systems are synchronized using the fs_copy command. This can be performed over the IP network or, if the source file system contains a large amount of data, by physically transporting the data to a remote site using disk or tape.

3. The addresses of all subsequent block modifications made to the source file system are used by replication to create one or more delta sets. The replication service creates a delta set by copying the modified blocks to the SavVol at the source site.

4. Replication transfers any available, complete delta sets which include the block addresses to the destination SavVol. During this time, the system tracks subsequent changes made to the source file system on the source site.

5. At the destination site, the playback service plays back any available, complete delta sets to the destination file system, which makes it consistent with the source file system.

Destination site

2

4

Source site

CNS-000762

Source storage unit

Celerra Network Server 1

Data Mover

Sourcefile

system

SourceSavVol

1

3

Destinationfile

system

Celerra Network Server 2

Data Mover

Destination storage unit

5

6

DestinationSavVol

Page 13: Celerra_Replcatr

13 of 172Version 5.6.47Using Celerra Replicator (V1)

6. The Data Mover at the destination site exports the read-only copy for content distribution, backup, and application testing. This optional step is done manually.

Activating the destination file system as read/write

If the source file system becomes unavailable, usually as the result of a disaster, you can make the destination file system read/write for local or remote scenarios. After the source site is available again, you can then restore replication to become read/write at the source site and read-only at the destination site. This three-stage process includes:

◆ Using the destination file system for production when the source file system is unavailable (failover).

◆ Resynchronizing the file systems.

◆ Restoring the replication process to its original state.

Replication failover

In this example, the source site has experienced a disaster and is unavailable. Failover ends the replication relationship between the source and destination file systems and changes the destination file system from read-only to read/write.

When failing over, the following actions occur:

1. The system stops replication and plays back the outstanding delta sets on the destination site to the destination file system according to the options specified by the user. The system can play back either all or none of the delta sets. The beginning of the failover process is shown in Figure 3 on page 14.

Page 14: Celerra_Replcatr

Using Celerra Replicator (V1)14 of 172 Version 5.6.47

Figure 3 Source site becomes unavailable

2. The system stops the playback service and the destination file system becomes read/write, as shown in Figure 4 on page 15. This illustration displays a remote replication scenario with the destination site as read/write. This state is possible in a local replication scenario as well.

Destination siteSource site

CNS-000763

Source storage unit

Celerra Network Server 1

DataMover

Sourcefile

system

SourceSavVol

Destinationfile

system

Celerra Network Server 2

DataMover

Destination storage unit

DestinationSavVol

Source site

Source storage unit

Celerra Network Server 1

DataMover

Sourcefile

system

SourceSavVol

Page 15: Celerra_Replcatr

15 of 172Version 5.6.47Using Celerra Replicator (V1)

Figure 4 Failover

3. The destination site can be enabled to allow read/write access to the destination file system from network clients (local or remote scenario). "After a failover or reversal" on page 78 provides more information. This optional step is done manually.

Note: If the source file system is online, it becomes read-only.

Destination siteSource site

CNS-000764

Source storage unit

Celerra Network Server 1

DataMover

Sourcefile

system

SourceSavVol

Destinationfile

system

Celerra Network Server 2

DataMover

Destination storage unit

DestinationSavVol

Source site

Source storage unit

Celerra Network Server 1

DataMover

Sourcefile

system

SourceSavVol

Page 16: Celerra_Replcatr

Using Celerra Replicator (V1)16 of 172 Version 5.6.47

Resynchronization

When the original source file system becomes available, the replication relationship can be reestablished. The fs_replicate -resync option is used to populate the source file system with the changes made to the destination file system while the source site was unavailable. This establishes the replication relationship in the reverse direction. The destination file system is read/write and the source file system is read-only, as shown in Figure 5 on page 16.

Figure 5 Resynchronization

Destination site

Sourcefile

system

Celerra Network Server 1

Data Mover

Source storage unit

Celerra Network Server 2

Data Mover

Destination storage unit

Source site

Destination SavVol

CNS-000761

SourceSavVol

Destinationfile

system

Page 17: Celerra_Replcatr

17 of 172Version 5.6.47Using Celerra Replicator (V1)

Replication reversal

In this example, the reversal is used after a failover and resynchronization to change the direction of the replication. The source site again accepts the source file system updates from the network clients, and the replication service transfers these updates to the destination site for playback to the destination file system, as shown in Figure 6 on page 17.

Figure 6 Replication reversal

Note: A reversal requires both sites to be available and results in no data loss. During the reversal phase, the source and destination file systems are set as read-only while the last updates are transferred.

"Recovering replication data" on page 61 further describes the replication reversal feature.

Communication between Celerra Network Servers

At the source and destination sites, you must build a trust relationship enabling HTTP communication between Celerra Network Servers. This trusted relationship is built upon a common passphrase set for both Celerra Network Servers. The 6- to 15-character passphrase is stored in clear text. It is used to generate a ticket for Celerra-to-Celerra communication.

Destination siteCelerra Network Server 1

DataMover

Source storage unit

Destinationfile

system

Celerra Network Server 2

DataMover

Destination storage unit

Source site

DestinationSavVol

CNS-000766

SourceSavVol

Sourcefile

system

Page 18: Celerra_Replcatr

Using Celerra Replicator (V1)18 of 172 Version 5.6.47

Note: You use the nas_cel -create command to establish the relationship between your Celerra Network Servers. For this command, ensure that the passphrase has at least 6 characters and no more than 15 characters. Remember that you must run the command on both Celerra Network Servers and use the same passphrase.

The time on the Data Movers involved in a replication relationship and the Control Stations at both sites must be synchronized with a maximum allowable skew of 10 minutes. Take into account time zones and daylight savings time, if applicable, when using the Network Time Protocol (NTP) to synchronize the time.

Configuring EMC Celerra Time Services offers more information.

To establish communication, first, you must have root privileges and each site must be active and configured for external communication. Table 1 on page 18 shows information about the source and destination sites used in these examples.

Second, there must be IP network connectivity between both Control Stations. Verify whether a relationship exists using the nas_cel -list command. If communication is established, go to Task 3: "Create SnapSure checkpoint of source file system" on page 42.

Note: This task is performed only for remote replication.

How resynchronization works

After a failover completes and the source site becomes operational, you can resume replication using the -resync option. When the file systems are resynchronizing, changes that occurred after the failover are copied to the source site and replication is started but the replication direction is reversed. The replication service is running on the destination site, and the playback service is running on the source site. If this resynchronization is successful, you need not perform a full file system copy.

Some reasons a resynchronization may not be possible are:

◆ You performed a failover but the system was unable to make the source file system read-only. When the source site became available, it continued to receive I/O to your source file system. If your replication service becomes inactive, you must abort replication. You should not continue to allow I/O to the original source file system.

◆ After you performed a failover, you decided to abort replication when the source site became available.

Table 1 Source and destination sites information

Site Celerra name IP address

[source_site] cs100 192.168.168.114

[destination_site] cs110 192.168.168.102

Page 19: Celerra_Replcatr

19 of 172Version 5.6.47Using Celerra Replicator (V1)

If the replication service is active and you receive one of the following error messages when attempting a resynchronization, you can run the -resync command again by specifying the autofullcopy=yes option:

Error 2242: <file_system_name> : replication is not active, incremental resync not possible. Abort and start replication or resync with autofullcopy=yes.

or

Resync copy failed. Incremental resync not possible. Abort and start replication or resync with autofullcopy=yes.

Note: If this incremental resynchronization fails, restarting replication using a full file system copy might take considerable time and resources. Plan carefully before using this option.

!CAUTION!Any data on the source file system not played back to the destination file system prior to the failover is permanently lost.

How suspend works

Suspend is an option that allows you to temporarily stop an active replication relationship and leave the replication in a condition that allows it to be restarted. Suspend, when used in conjunction with the restart option, allows you to temporarily stop replication, perform some action, and then restart the replication relationship using an incremental rather than a full data copy.

After suspending a replication relationship, you can:

◆ Change the replication SavVol size. During the course of using replication, the size of a SavVol may need changing because:

• The SavVol is too large and you want to reclaim the unused disk space.

• The SavVol is too small, which activates flow control.

◆ Mount the replication source or destination file system on a different Data Mover.

◆ Change the IP addresses or interfaces the replication is using.

When you are restarting a replication relationship, you can specify a source interface or allow the system to select it:

• If you specify an interface for the source site, replication uses that interface until the user changes it.

• If you allow the system to select the interface, the interface can change to keep the replication relationship running.

For example, if the network interface currently in use becomes unavailable, the system attempts to select another interface. If it finds one, the replication relationship continues to function.

The destination interface, regardless of how it is selected, is unchanged by the system.

Page 20: Celerra_Replcatr

Using Celerra Replicator (V1)20 of 172 Version 5.6.47

When suspending a replication relationship, the system:

◆ Ensures all delta sets are transferred to the destination site.

◆ Plays back all outstanding delta sets.

◆ Creates a checkpoint on the source site, which is used to restart replication.

Note: The suspend checkpoint is named root_suspend_ckpt_xxx_1, where xxx is the ID of the suspended replication relationship destination and 1 represents Celerra ID of a remote replication relationship. No number appears for a local replication relationship.

The replication and playback services are no longer running after the suspend action is complete.

How replication relationship restarts

The restart option allows you to restart a replication relationship by using an incremental rather than a full data copy. Use the restart option when a replication relationship is:

◆ Suspended

◆ Out-of-sync

Restart a suspended replication relationship

After you suspend a replication relationship using the -suspend option, only the -restart option can restart it. This command verifies that the replication is in a condition to allow a restart. It begins the process with an incremental copy using a checkpoint of the source file system created when the replication suspended. You must include this checkpoint when determining the maximum number of checkpoints per file system used with replication.

Note: Before you restart a replication, make sure that all checkpoints are mounted. Otherwise, a full data copy will be initiated instead of an incremental copy.

If you are using this procedure to:

◆ Increase the size of the replication SavVol, ensure that you specify the new SavVol size using the savsize=<newsize> option.

◆ Change interfaces or IP addresses, specify them when you restart the replication relationship.

When replication is restarted, default values are used for the replication policies. For example, high water mark and timeout are set to 600. Specify new policies when you restart replication using -option <options>.

Out-of-sync replication relationship

The source and destination file systems can become out-of-sync because:

◆ Network connectivity is lost.

◆ Incoming write rate is greater than the delta-set replay rate on the destination file system.

Page 21: Celerra_Replcatr

21 of 172Version 5.6.47Using Celerra Replicator (V1)

This causes the following:

1. Delta sets accumulate on the source site until the SavVol is filled.

2. Changes are logged in memory until they can no longer be tracked.

3. The file systems fall out-of-sync if no source policy is specified.

Prevent an out-of-synchronization condition by specifying a source policy for a replication relationship. Either stop accepting writes or stop all client access to the file system.

Reestablishing the replication relationship

When the network connection is reestablished or the source file system is available again, you can restart replication. To restart, you must have usable checkpoints of the source file system. When restarting, all previous configuration parameters are maintained and cannot be changed during the restart.

When restarting an out-of-sync replication relationship, the system:

◆ Determines the optimal checkpoint that meets the criterion of having a delta number less than the delta-set number of the destination file system.

◆ Aborts the replication relationship.

◆ Restarts replication using the original configuration information.

◆ Performs an incremental copy of the file system using the appropriate checkpoint.

If no valid checkpoint is available, you must abort and reestablish the replication relationship. This means a complete (not an incremental) copy of the file system must be done.

Checkpoint availability

To ensure that a valid checkpoint is available to restart an out-of-sync replication relationship, create two checkpoints of the source file system. These restartable checkpoints, named <source_fs_name>_repl_restart_1 and <source_fs_name>_repl_restart_2, are automatically refreshed by a system CRON job that runs every hour at 25 minutes after the hour. When a replication relationship is out-of-sync, these checkpoints are available for the restart process. You must include these two checkpoints when determining the maximum number of checkpoints per file system used with replication.

Verify that these checkpoints are being refreshed by occasionally checking your server log (server_log) file. If the system is not refreshing your checkpoints regularly, call your service provider. The following is a sample message for a successful refresh, where <fs_name> is the file system name:

Dec 29 07:29:23 2004 NASDB:7:13 refresh scheduled replication restart ckpt <fs_name>_repl_restart_1 succeeded.

Page 22: Celerra_Replcatr

Using Celerra Replicator (V1)22 of 172 Version 5.6.47

System requirements for Celerra Replicator

This section details Celerra Network Server software, hardware, network, and storage settings to use Celerra Replicator as described in this technical module.

Local replication

Table 2 on page 22 lists the system requirements for local replication.

Remote replication

Table 3 on page 22 lists the system requirements for remote replication.

Table 2 Local replication system requirements

Software• Celerra Network Server version 5.6.• Licenses for Celerra Replicator and SnapSure.

Hardware

• One Celerra-storage (EMC Symmetrix® or EMC CLARiiON®) pair.• CNS-14 or CFS-14 requires a minimum of a 510 Data Mover or later.• To provide a graceful shutdown in an electrical power loss, Celerra Network

Server and the storage array need to have Uninterruptible Power Supply (UPS) protection. If this is not provided, replication becomes inactive.

Network • IP addresses configured for the primary and secondary Data Movers.

Storage• Sufficient storage space available for the source and destination file systems.• Sufficient SavVol space available for use by Celerra Replicator and

SnapSure.

Table 3 Remote replication system requirements (page 1 of 2)

Software

• Celerra Network Server version 5.6 with the same Celerra version on the source and destination Celerra Network Servers.

Note: For limited management capability and no fs_copy support: Celerra Network Server version 5.5.39.2 on the source and 5.6.47 on the destination Celerra Network Server.

• Licenses for Celerra Replicator and SnapSure.

Hardware

• Minimum of two Celerra-storage (Symmetrix or CLARiiON) pairs.• CNS-14 and CFS-14 require a minimum of one 510 Data Mover or later for

each Celerra-storage pair.• To provide a graceful shutdown in an electrical power loss, Celerra Network

Server and the storage array need to have UPS protection. If this is not provided, replication becomes inactive.

Page 23: Celerra_Replcatr

23 of 172Version 5.6.47Using Celerra Replicator (V1)

Consult "Configuration considerations" on page 35 for more information.

Network

• IP addresses configured for the source and destination Data Movers (ports 8888 and 8887 used by replication for transferring data and internal operations—contact Customer Service to change this setting).

• HTTPS connection between the Control Station on the source site and the Control Station on the destination site (port 443—cannot be changed).

• Internet Control Message Protocol (ICMP) ensures that a destination Celerra Network Server is accessible from a source Celerra Network Server. The ICMP protocol reports errors and provides control data about IP packet processing.

Storage• Sufficient storage space available for the source and destination file systems.• Sufficient SavVol space available for use by Celerra Replicator and

SnapSure.

Table 3 Remote replication system requirements (page 2 of 2)

Page 24: Celerra_Replcatr

Using Celerra Replicator (V1)24 of 172 Version 5.6.47

Upgrading from previous Celerra Network Server versions

This section describes the upgrade process when upgrading from a previous version of Celerra Network Server to version 5.6 when using Celerra Replicator (V1):

◆ Upgrade from a version earlier than 5.5.39.2

◆ Upgrade from version 5.5.39.2 or later (with out-of-family support)

Upgrade from a version earlier than 5.5.39.2

When upgrading from a previous version of Celerra Network Server that is earlier than 5.5.39.2, to 5.6, note the following:

◆ Local replication

If you are upgrading a local replication, stop the playback service for all the destination file systems by setting the timeout and high water mark to 0. Upgrade Celerra Network Server, reset the timeout, and high water marks for the destination file systems to their previous settings.

◆ Remote replication:

• If you are upgrading where only destination file systems are on the remote system, upgrade Celerra Network Server at the destination site first, and then upgrade the source site.

• You cannot upgrade a destination Celerra that is the target of a Celerra running a 5.5 version earlier than 5.5.39.2.

• You cannot upgrade a Celerra running a version earlier than 5.5.39.2 that is the source of an active replication session if the destination is also running a 5.5 version.

• You cannot upgrade a Celerra to 5.6 if there are bi-directional sessions targeting a 5.5 Celerra as a destination.

◆ General replication considerations:

• Do not perform replication operations while upgrading (for example,-failover, -resync, -reverse, -restart, -refresh).

• When upgrading the source and destination Celerra Network Servers from version 5.5 to 5.6, replication sessions cannot be administered until both sides have been upgraded to 5.6.

Page 25: Celerra_Replcatr

25 of 172Version 5.6.47Using Celerra Replicator (V1)

• Before starting an upgrade, ensure that no failed over or suspended replication relationships are present. If you upgrade with replication relationships in this state, these relationships will be unusable when the upgrade completes.

• Introduced in version 5.4, SnapSure checkpoints use a new pageable blockmap structure. Checkpoints used for replication are in this format. Using SnapSure on EMC Celerra describes the pageable blockmap structure in depth.

• Replication policies for time-out interval and high water mark are as follows:

– Time out: Acceptable timeout values are 0 or greater than 60 seconds up to a limit of 24 hours.

– High water mark: The high water mark maximum value is 8000 MB. The value should not exceed the size of the SavVol.

Note: You do not have to abort replication when upgrading Celerra Network Server version.

Upgrade from Celerra Network Server version 5.5.39.2 or later

Introduced in version 5.5.39.2, out-of-family replication support is recommended for customers in a multi Celerra, edge-to-core environment who want to migrate to NAS version 5.6 but want to upgrade their Celerras at different times without interrupting data transfer.

In an edge-to-core configuration, where there are multiple source Celerras replicating to one destination Celerra, you upgrade the destination (core) Celerra to NAS version 5.6 first and then upgrade each of the source (edge) Celerras when appropriate.

Out-of-family replication is not intended for use over an extended period of time as there is limited replication management capability while in this environment. For example, you cannot start a new replication, restart a suspended or inactive replication, resync a failed-over replication, suspend or reverse a replication, or perform a copy (full or differential) of a file system. Table 4 on page 26 provides the support matrix for replication commands when in an out-of-family configuration. All management commands are supported after both sides of the replication session have been upgraded to the same NAS code family version.

When upgrading from version 5.5.39.2 or later of Celerra Network Server to 5.6, note the following:

◆ Out-of-family replication support requires NAS version 5.5.39.2 or later on the source Celerra and 5.6.47 or later on the destination Celerra.

◆ You cannot upgrade a 5.5.39.2 Celerra that is the source of an active replication session if the destination is also running version 5.5.39.2.

◆ You cannot upgrade a Celerra to 5.6 if there are bi-directional sessions targeting a 5.5 Celerra as a destination.

Page 26: Celerra_Replcatr

Using Celerra Replicator (V1)26 of 172 Version 5.6.47

◆ Out-of-family replication is unidirectional from a 5.5.39.2 source to a destination running 5.6.47 or later. Replication from source 5.6 to destination 5.5 is not supported.

◆ Does not support fs_copy.

◆ After upgrade, there is limited replication management support for Celerra Replicator (V1) sessions running with different NAS code family versions (source running 5.5.39.2 and destination running 5.6). Table 4 on page 26 provides the support matrix of replication management commands when in an out-of-family configuration.

Table 4 Out-of-family replication command support matrix (page 1 of 2)

CommandReplication state

Active Inactive Suspended Failed-over Not configured

Abort on source Allowed Allowed NA Allowed NA

Abort on destination Allowed Allowed NA NA NA

Abort on both Not allowed Not allowed NA Not allowed NA

Refresh on source Allowed NA NA Allowed NA

Refresh on destination

Allowed Allowed NA NA NA

Refresh on both Not allowed NA NA Not allowed NA

Modify on source Allowed Allowed NA Allowed NA

Modify on destination Allowed Allowed NA NA NA

Modify on both Not allowed Not allowed NA Not allowed NA

Failover on destination (default | Now | Sync)

Allowed (default | now)

Not Allowed (sync)

Allowed (default | now)

Not Allowed (sync)

NA NA NA

Resync on destination

NA NA NA Not allowed NA

Reverse on source Not allowed Not allowed NA NA NA

Restart on source NA Not allowed Not allowed NA NA

Suspend on source Not allowed Not allowed NA NA NA

List from source Allowed Allowed NA Allowed NA

List from destination Allowed Allowed NA NA NA

Page 27: Celerra_Replcatr

27 of 172Version 5.6.47Using Celerra Replicator (V1)

Prerequisites

Before upgrading, make sure that the Celerra to be updated:

◆ Is running version 5.5.39.2 or later.

◆ If Celerra to be upgraded is a destination, make sure that the source hosting the active replication sessions is running version 5.5.39.2 or later.

◆ If Celerra to be upgraded is a source, make sure that the destination hosting the active replication sessions is running version 5.6.47 or later.

◆ Has no fs_copy sessions running.

◆ Is running under minimum load. If there is a high rate of I/O during upgrade it may cause replications to become inactive.

◆ Is not hosting both the source and destination sides of active replications running version 5.5. If bi-directional sessions exist on the Celerra to be upgraded, do the following:

a. Suspend all replication sessions in one direction. (Either the sessions running from A to B or the sessions running from B to A.)

b. Upgrade the Celerra that is hosting only the destination side of the active replication sessions.

c. Data transfer will continue but with limited replication management capability.

d. Upgrade the source Celerra.

e. Restart all the suspended replication sessions.

Procedure

To upgrade to Celerra Network Server version 5.6.47:

Info from source Allowed Allowed NA Allowed NA

Info from destination Allowed Allowed NA NA NA

Start from source NA Not allowed NA NA Not allowed

fs_copy from source NA NA NA NA Not allowed

Table 4 Out-of-family replication command support matrix (page 2 of 2)

CommandReplication state

Active Inactive Suspended Failed-over Not configured

Step Action

1. Upgrade the destination Celerra from NAS code version 5.5 to 5.5.39.2.

2. Upgrade all of the source Celerras from NAS code version 5.5 to 5.5.39.2.

Page 28: Celerra_Replcatr

Using Celerra Replicator (V1)28 of 172 Version 5.6.47

3. Upgrade the destination Celerra from NAS code version 5.5.39.2 to 5.6.47.

Data continues to transfer between the source and destination sites, but there is limited replication management capability until you upgrade the source Celerra to 5.6.47. For example, you cannot start a new replication, restart a suspended or inactive replication, resync a failed-over replication, suspend or reverse a replication, or perform a copy (full or differential) of a file system. Table 4 on page 26 provides the out-of-family replication command support matrix.

4. Upgrade the source Celerra from 5.5.39.2 to 5.6.47.

5. Repeat step 4 for all of the source Celerras to be upgraded.

Step Action

Page 29: Celerra_Replcatr

29 of 172Version 5.6.47Using Celerra Replicator (V1)

Planning considerations for Celerra Replicator

Before you use Celerra Replicator:

◆ Review replication policies

◆ Determine SavVol size

◆ Determine the number of replications per Data Mover

◆ Review the configuration considerations

Replication policies

Most replication policies can be established for one replication relationship (using the fs_replicate command) or all replication relationships on a Data Mover by setting a parameter. Celerra Replicator has policies to:

◆ Control delta-set generation using a time-out interval and high water mark.

◆ Control how to handle data if network connectivity is lost (flow control). "Celerra Replicator flow control" on page 31 describes this action.

◆ Set the maximum IP bandwidth size used by a replication session. "Set bandwidth size" on page 109 details this policy.

◆ Set the data amount to be sent across the IP network before an acknowledgment is required from the receiving side. This is controlled by a parameter for the TCP window size (tcpwindow). "Accommodating network concerns" on page 127.

IP Alias with IP replication

Celerra Network Server versions 5.5.28.1 and later support IP Alias with IP replication. All restrictions on Control Station failover also apply to IP Alias with IP replication configurations. The following guidelines apply to this feature:

◆ When using IP replication for the first time, or on new systems, configure IP Alias first, and use IP Alias in the -ip <ipaddr> option of nas_cel -create command.

◆ For existing systems with existing IP replication sessions, the current slot_0 IP address (primary Control Station IP address) must be used. For example:

nas_config -IPalias -create 0 -> Do you want slot_0 IP address as your alias [yes or no] yes

◆ If the Control Station fails over while IP replication is running, the IP replication command (fs_replicate) might need to be re-executed manually. Check logs (/nas/log/cmd_log*, server_log command output etc.) to determine how to proceed. Keep the fs_replicate command output for resync, suspend, restart, failover, and reverse options in case of failure, then execute the rest of steps based on the instruction in command output.

◆ When IP Alias is deleted using the nas_config -IPalias -delete command, the IP address of the primary or the secondary Control Station is not changed. Changes to the IP address of the primary or the secondary Control Station must be done separately. IP replication depends on communication between the

Page 30: Celerra_Replcatr

Using Celerra Replicator (V1)30 of 172 Version 5.6.47

source Control Station and the destination Control Station. When IP Alias is used for IP replication, deleting the IP Alias breaks the communication. The IP address which was used as the IP Alias must be restored on the primary Control Station to restore the communication.

◆ While performing a Celerra code upgrade using ssh or telnet, do not use an IP Alias in the ssh or telnet to log in to the Control Station.

Control delta-set generation

A delta set contains the block modifications made to the source file system and is used by the replication service to periodically synchronize the destination file system with the source file system. The amount of information within the delta set is based on the source file system’s activity and how you set the replication policies, time out, and high water mark values. The minimum delta-set size is 128 MB.

The replication service is triggered by either the time-out policy or the high water mark policy, whichever is reached first. However, when the maximum delta-set size is reached (8 GB), a new delta set is generated, regardless of the time-out policy:

◆ Time-out policy (where the system either generates or plays back a delta set):

• Source site: The interval at which the replication service automatically generates a delta set (for example, every 1200 seconds).

• Destination site: The interval at which the playback service automatically plays back all available delta sets to the destination file system (for example, every 600 seconds).

At both sites, the default time-out value is 600 seconds. Acceptable time-out values are 0 or greater than 60 seconds up to a limit of 24 hours. A value of 0, indicating there is never a timeout, pauses the replication activities for this policy.

◆ High water mark:

• Source site: The size in MB of the file system changes accumulated since the last delta set, at which the replication service automatically creates a delta set on the SavVol. For example, when the amount of change reaches 1200 MB in size, a delta set is generated.

• Destination site: The size in MB of the delta sets present on the destination SavVol, at which the replication service automatically plays back all available delta sets to the destination file system. For example, when the amount of change reaches 600 MB in size, playback occurs.

At both sites, the default high water mark value is 600 MB. Acceptable high water mark values are 0 or greater than 30 MB up to a maximum value of 8 GB. This value should not exceed the size of the SavVol. A value of 0 pauses the replication activities and disables this policy.

Note: A delta set may not be generated or copied if a flow control is triggered. "Celerra Replicator flow control" on page 31 provides further information.

Page 31: Celerra_Replcatr

31 of 172Version 5.6.47Using Celerra Replicator (V1)

How time-out and high water mark policies work

How time-out and high water mark policies work discusses how these replication policies work for the source and destination file systems.

Replication policies for the source file system and when one of these replication policies is triggered, the following occurs:

1. The replication service automatically generates a delta set from the accumulated changes to the file system and stores it in the SavVol. Each delta set is recorded and processed per replication policy trigger. Each delta set contains a set of changes to the source file system that occur since the creation of the previous delta set.

2. Replication tracks all subsequent changes made to the source file system.

3. The delta sets are transferred from the SavVol on the source site to the SavVol on the destination site (for remote replication), and updates the destination file system through the replication playback service (for local and remote replication).

4. The replication service waits for the next event to create the next delta set.

Replication policies for the destination file system:

◆ The playback service continually polls the SavVol on the destination site to play back each delta set to the destination file system, synchronizing it with the source file system. This playback rate is based on the specified replication policy. After the delta set is copied to the destination file system, the next delta set is processed.

◆ For optimization, delta sets available before a trigger is reached are merged into one active playback session.

Celerra Replicator flow control

Celerra Replicator flow control activates when the playback rate on the destination site is too slow, when network connectivity is lost, or when the write rate from the source site is too fast.

Celerra Replicator activates flow control in an attempt to allow the network to catch up. In most cases, Celerra Replicator should not be in a flow control state. Flow control is activated:

◆ If a delta set that has not been replayed on the destination file system is about to be overwritten by a newer delta set, the destination file system temporarily holds the data flow until the delta set is replayed. This happens when the delta-set playback rate for the destination file system is too slow to handle the source file system updates.

◆ When a delta set cannot be transferred from the SavVol on the source site to the SavVol on the destination site (for example, the network is unavailable), the replication service stops the data flow to the destination site. During this time, Celerra Replicator tracks the source file system modifications, and continually retries connectivity to the destination site. If the network is down, a message is sent to the system log on the source site.

◆ When the SavVol at the source site is full, the replication service suspends copying the modified blocks to the SavVol. This prevents the overwriting of delta

Page 32: Celerra_Replcatr

Using Celerra Replicator (V1)32 of 172 Version 5.6.47

sets not yet transferred to the SavVol on the destination site. During this time, modifications to the source file system are still tracked, but are not copied to the SavVol until it has available space.

Celerra Network Server tracks changes in memory only until these changes represent a delta set the same size as the Celerra Replicator SavVol. If the system can no longer track changes in memory, the system behaves in one of three ways, as explained in Table 5 on page 32.

As shown in Table 5 on page 32, these policies can be set for one replication session or for all replication sessions on a Data Mover. If there is a conflict between these two policies, the one defined for a single replication session takes precedence.

You can set up alerts to notify you when these events occur. "Events for Celerra Replicator" on page 121 details Celerra Replicator events, and Configuring EMC Celerra Events and Notifications describes how to use them.

Table 5 Replication flow-control options

BehaviorHow to specify for a single replication session

How to specify for all replication sessions on a Data Mover

Temporarily halts all I/Os to the source file system until sufficient space is available on the source SavVol. During this time, the file system is inaccessible to network clients.

When space is available on the Celerra Replicator SavVol, the source file system is mounted and begins accepting I/Os.

Set an option using fs_replicate -modify-option autofreeze=yes. "How time-out and high water mark policies work" on page 31 describes this policy.

Set the VRPL freeze parameter as described in the EMC Celerra Network Server Parameters Guide.

Temporarily stops writing data to the source file system by mounting it read-only. Users still have read-only access to the source file system.

When space becomes available on the Celerra Replicator SavVol, the source file system is remounted read/write and begins accepting all I/Os to the source file system.

Set an option using fs_replicate -modify-option autoro=yes. "How time-out and high water mark policies work" on page 31 describes this policy.

Set the VRPL read-only parameter as described in the EMC Celerra Network Server Parameters Guide.

Allows the replication service to fall out of sync.

"Restarting a replication relationship" on page 89 describes how to restart replication after it has fallen out of sync.

Not applicable.

Page 33: Celerra_Replcatr

33 of 172Version 5.6.47Using Celerra Replicator (V1)

SavVol size requirements for remote replication

Replication SavVols store changes that occurred on the source file system not yet replayed to the destination file system. Celerra Replicator requires a SavVol at source and destination sites. The minimum SavVol size is 1 GB and the maximum is 500 GB. Determine the size of these SavVols based on the following criteria:

◆ Investigate the source file system size, update ratio of the source file system per day, and the WAN network bandwidth for the source/destination file systems between the source and destination Celerra Network Servers. Use nas_fs -size to calculate the SavVol size and nas_disk -list to find the entire file system size. The EMC Celerra Network Server Command Reference Manual provides more command details.

◆ Evaluate the risk tolerance to network outages. For example, if the network experiences long outages, such as two days, ensure that the SavVol on the source site will allow capturing two days of delta sets.

◆ If performing replication on multiple file systems per Data Mover, consider the available network bandwidth per file system.

◆ Determine whether the network bandwidth is sufficient to transfer the changed data from the source to destination file system. If the rate of change on the source is continuously greater than available network bandwidth, the replication service will not transfer data quickly enough, eventually becoming inactive.

You may also contact EMC Customer Service or read the E-Lab Interoperability Navigator on Powerlink to size the SavVols.

Change replication SavVol requirements

In most cases, the default size of the SavVol is sufficient (10 percent of the source file system size). To accommodate a large network outage or allow for brief periods where the incoming change rate significantly exceeds the ability of the network to send changes to the destination site or both, consider increasing your replication SavVol size.

For example, a 50 GB SavVol is sufficient for a 500 GB file system that incurs 20 GB of change per day to cover approximately two and one-half days of network outage without any flow-control events. But to cover a longer outage period, you can enlarge the SavVol.

Page 34: Celerra_Replcatr

Using Celerra Replicator (V1)34 of 172 Version 5.6.47

Table 6 on page 34 describes changing the replication SavVol size and explains when to use each method.

Determine the number of replications per Data Mover

Determine the maximum number of replication sessions per Data Mover based on your configuration (such as the WAN network bandwidth, delta-set update ratio, and the production I/O workload). This number is also affected by whether you are running SnapSure and Celerra Replicator on the same Data Mover. Both of these applications share the available memory on a Data Mover.

For all configurations, there is an upper limit to the number of replications allowed per Data Mover. E-Lab Interoperability Navigator on Powerlink details the current number of replications allowed per Data Mover.

Note: If you plan to run loopback replications, remember that each loopback replication counts as two replication sessions because each session encapsulates outgoing and incoming replications.

To learn the number of replication sessions per Data Mover:

1. Determine the SavVol size of each replication, as described in "SavVol size requirements for remote replication" on page 33.

2. Verify that the total storage on a Data Mover (including any source and destination file systems and associated SavVols) does not exceed the guidelines for that Data Mover. These guidelines are detailed in the E-Lab

Interoperability Navigator on Powerlink.

3. Celerra Replicator should be transferring delta sets faster than it creates them. Verify that the delta sets per Data Mover can be transferred to the destination

Table 6 Changing Celerra Replicator SavVol size

When to use What to do Procedure

Before starting any replication processing.

Change default size of each Celerra Replicator SavVol from 10% of the source file system.

By default, the system allocates 10% of the size of the source file system for the replication SavVol on the source and destination sites.

"Change the Celerra Replicator SavVol default size" on page 123

At the start of a replication instance.

Control SavVol size for a file system by specifying a specific SavVol size. Use savsize option for fs_replicate -start.

Task 6: "Begin replication" on page 48

After replication is running. Revise SavVol size to meet your changing requirements.

"Suspend a replication relationship" on page 81

Page 35: Celerra_Replcatr

35 of 172Version 5.6.47Using Celerra Replicator (V1)

site with the available WAN network bandwidth. An active flow-control condition may be indicative of this situation.

Note: To provide a stable network transfer rate for delta sets, it is strongly recommended that you configure a dedicated network port for Data Mover transfers.

4. Verify that the Data Mover can handle all replication sessions and production I/Os. You can also monitor memory usage and CPU usage using the server_sysstat command. This command shows total memory utilization, not just Celerra Replicator and SnapSure memory usage.

Note: Use Celerra Manager to monitor memory and CPU usage by creating a new notification on Celerras > Notifications > Data Mover Load tab.

Contact EMC Customer Service for additional advice.

Configuration considerations

Before setting the replication policy triggers, consider the following:

◆ To avoid the source and destination file systems from becoming out-of-sync, do not allow the replication service to create delta sets significantly faster than it can copy them to the destination file system. Set the delta-set creation replication policy to a higher number (for example, 1200 seconds) than the delta-set playback number (for example, 600 seconds). The replication policies you establish for creating and replaying delta sets depend on the size and number of transactions processed on the source file system.

◆ Determine if the network bandwidth can effectively transport the production change data generated at the source site to the destination site.

◆ During the delta-set playback on the destination file system, network clients can access the destination file system. However, at the beginning of the delta-set playback for CIFS (Common Internet File Service) clients, there is a temporary freeze/thaw period that may cause a network disconnect. As a result, do not set the replication policy to a low number because this reduces the availability of the destination file system.

To eliminate this freeze/thaw period, create a checkpoint of the destination file system and mount it for client access at the destination site. However, this checkpoint will not contain the most up-to-date production data.

◆ Carefully evaluate the infrastructure of the destination site by reviewing items such as:

• Subnet addresses

• Unicode configuration

• Availability of name resolution services; for example, WINS, DNS, and NIS

• Availability of WINS/PDC/BDC/DC in the correct Microsoft Windows NT, or Windows Server domain

• Share names

Page 36: Celerra_Replcatr

Using Celerra Replicator (V1)36 of 172 Version 5.6.47

• Availability of user mapping (for example, using EMC Usermapper for Celerra systems). The CIFS environment requires more preparation to set up a remote configuration because of the higher demands on its infrastructure than the network file system (NFS) environment (for example, authentication is handled by the domain controller). For the CIFS environment, you must perform mappings between the usernames/groups and UIDs/GIDs with EMC usermapper or local group/password files on the Data Movers.

Note: Replicating EMC Celerra CIFS Environments (V1) describes configuration considerations in depth.

◆ Local groups are not supported on replicated file systems unless you use VDMs. Replicating EMC Celerra CIFS Environments (V1) describes this consideration more fully.

◆ The replication SavVol for the delta sets must be large enough to store and process all the delta-set write I/Os, and the SnapSure SavVol for the checkpoints must be able to store all the source file system block changes for the initial synchronization.

◆ The destination file system can only be mounted on one Data Mover, even though it is read-only.

◆ At the application level, as well as the operating system level, some applications might have limitations on the read-only destination file system due to caching and locking.

◆ If you are planning on enabling international character sets (Unicode) on your source and destination sites, you must first set up translation files on both sites before starting Unicode conversion on the source site. Using International Character Sets with EMC Celerra covers this consideration.

◆ Celerra FileMover feature supports replicated file systems. This is described in Using EMC Celerra FileMover.

◆ Celerra File-Level Retention Capability supports replicated file systems. Using File-Level Retention on EMC Celerra provides additional configuration information.

Page 37: Celerra_Replcatr

37 of 172Version 5.6.47Using Celerra Replicator (V1)

User interface choices for Celerra Replicator

Celerra Network Server offers flexibility in managing networked storage based on your support environment and interface preferences. This technical module describes how to set up and manage replication using the command line interface (CLI). You can also perform most tasks using Celerra Manager — Basic Edition.

The following documents provide additional information about managing Celerra:

◆ Getting Started with Celerra details user interface choices.

◆ Learning about EMC Celerra on the EMC Celerra Network Server Documentation CD and the application’s online help system describe each application’s capabilities.

◆ The EMC Celerra Network Server Release Notes provide additional, late-breaking information about Celerra management applications.

Page 38: Celerra_Replcatr

Using Celerra Replicator (V1)38 of 172 Version 5.6.47

Roadmap for Celerra Replicator

This section lists the tasks for configuring and managing Celerra Replicator.

Celerra Replicator configuration tasks:

1. "Initiating replication" on page 39

2. "Recovering replication data" on page 61

Celerra Replicator management tasks:

◆ "Abort Celerra Replicator" on page 79

◆ "Suspend a replication relationship" on page 81

◆ "Restarting a replication relationship" on page 89

◆ "Extending the size of a file system" on page 98

◆ "Resetting replication policy" on page 105

◆ "Reverse the direction of a replication relationship" on page 111

◆ "Monitor replication" on page 114

◆ "Checking playback service and outstanding delta sets" on page 115

◆ "Events for Celerra Replicator" on page 121

◆ "Change the Celerra Replicator SavVol default size" on page 123

◆ "Change the passphrase between Celerra Network Servers" on page 124

◆ "Managing and avoiding IP replication problems" on page 125

◆ "Transporting replication data using disk or tape" on page 139

◆ "Setting up the CLARiiON disk array" on page 147

Page 39: Celerra_Replcatr

39 of 172Version 5.6.47Using Celerra Replicator (V1)

Initiating replication

In most cases, you will have a functioning NFS or CIFS environment before you use Celerra Replicator. If you do not, ensure that you set up the following on the source and destination Celerra Network Servers before using Celerra Replicator:

◆ Establish the IP infrastructure.

◆ Establish name service. Read Configuring EMC Celerra Naming Services for more information about establishing name service.

◆ Synchronize the time between the Data Movers and Control Stations involved in the replication relationship. The maximum allowable time skew is 10 minutes. Read Configuring EMC Celerra Time Services for more information.

◆ Establish user mappings. Read Configuring EMC Celerra User Mapping for more information on establishing user mappings.

The process of setting up a local or remote replication relationship assumes the following:

◆ The source file system is created and mounted as read/write on a Data Mover.

◆ The destination file system is not created.

◆ The Celerra Network Server version of the destination site must be same as the version of the source site.

Note: The communication between Celerra Control Stations uses HTTPS.

When using remote replication, it is useful to create SnapSure checkpoints of the source file system to restart the source and destination sites if they fall out-of-sync. After you create the checkpoints, the system automatically keeps them up to date, as described in "Out-of-sync replication relationship" on page 20.

To set up replication you have to complete the following tasks:

1. "Establish communication" on page 40

2. "Verify communication" on page 40

3. "Create SnapSure checkpoint of source file system" on page 42

4. "Create the destination file system" on page 45

5. "Copy checkpoint to the destination file system" on page 45

6. "Begin replication" on page 48

7. "Create a second checkpoint of the source file system" on page 50

8. "Copy incremental changes" on page 52

9. "Verify file system conversion" on page 54

10. "Check replication status" on page 55

11. "Create restartable checkpoints" on page 59

Page 40: Celerra_Replcatr

Using Celerra Replicator (V1)40 of 172 Version 5.6.47

Note: The first two tasks are used for remote replication and do not apply to setting up a local replication relationship. The commands for setting up a local replication relationship are included for your reference. The output reflects a remote replication.

Task 1: Establish communication

To establish communication:

Task 2: Verify communication

This task is performed only for remote replication. To verify communication:

◆ "Verify communication at the source site" on page 41

◆ "Verify communication at the destination site" on page 41

◆ "View passphrase" on page 42

Action

To establish a trusted relationship at each site, logged in as root, use this command syntax:# nas_cel -create [-name <cel_name>] <ip> -passphrase <passphrase>

where:<cel_name> = name of the remote (destination) Celerra Network Server in the configuration<ip> = IP address of the remote Control Station in slot 0 <passphrase> = secure passphrase used for the connection, which must have 6- to 15- characters and be the same on both sides of the connection

Example:

To set up a trust relationship, type the following commands at both sites:[source_site]# nas_cel -create eng25271 -ip 172.24.252.71 -passphrase nasadmin[destination_site]# nas_cel -create eng25246 -ip 172.24.252.46 -passphrase nasadmin

Note: If you need to change the passphrase later, follow the procedure described in "Change the passphrase between Celerra Network Servers" on page 124.

Output

From source site eng25271, to set up relationship with destination site eng25246:

operation in progress (not interruptible)...id = 3name = eng25271owner = 0device =channel =net_path = 172.24.252.71celerra_id = 0001901003890010passphrase = nasadmin

Page 41: Celerra_Replcatr

41 of 172Version 5.6.47Using Celerra Replicator (V1)

Verify communication at the source site

At the source site, check whether Celerra Network Servers can communicate with one another.

Verify communication at the destination site

At the destination site, check whether Celerra Network Servers can communicate with one another.

Action

To verify that the source and destination sites can communicate with each other, type the command at each site: [source_site]$ nas_cel -list

Output

id name owner mount_dev channel net_path CMU0 cs100 0 192.168.168.114 APM0003400006800001 eng168123 201 xxx.xxx.xxx.xxx APM0004370489400003 eng16853 501 xxx.xxx.xxx.xxx 00018350173700005 cs110 503 192.168.168.102 APM000446038450000

Note

The sample output shows the source site can communicate with the destination site, cs110.

Action

To verify that the source and destination sites can communicate with each other, at each site, type:[destination_site]$ nas_cel -list

Output

id name owner mount_dev channel net_path CMU0 cs110 0 192.168.168.102 APM0004460384500002 cs100 501 192.168.168.114 APM000340000680000

Note

The sample output shows the destination site can communicate with the source site, cs100.

Page 42: Celerra_Replcatr

Using Celerra Replicator (V1)42 of 172 Version 5.6.47

View passphrase

The 6- to 15-character passphrase is used to authenticate with a remote Celerra Network Server.

Task 3: Create SnapSure checkpoint of source file system

A SnapSure checkpoint is used as the baseline of data to copy to the destination file system. Copying this baseline data from the source to the destination site over the IP network can be a time-consuming process.

When using remote replication, you can use an alternate method to copy the initial checkpoint of the source file system. You can back it up to a disk array or tape drive and transport it to the destination site. To use this alternate method, go to "Transporting replication data using disk or tape" on page 139 to continue with this procedure. Using SnapSure on EMC Celerra provides details on SnapSure checkpoints.

!CAUTION!◆ When creating checkpoints, be careful not to exceed your system’s limit. Celerra

permits 96 checkpoints per PFS, regardless of whether the PFS is replicated, for all systems except the Model 510 Data Mover (which permits 32 checkpoints with PFS replication and 64 checkpoints without). This limit counts existing checkpoints, or those already created in a schedule and might count two restartable checkpoints as well as a third checkpoint created by certain replication operations on either the PFS or SFS.

◆ If you are at the limit, delete existing checkpoints to create space for newer checkpoints, or do not create new checkpoints if existing ones are more

Action

To view the passphrase of a Celerra Network Server, use this command syntax:$ nas_cel -info id=<cel_id>

where:<cel_id> = Celerra ID

Note: Celerra ID is assigned automatically. To view this ID of a remote system, use the nas_cel -list command. You can also use the hostname.

Example:

To view the passphrase of the Celerra system, type:$ nas_cel -info id=5

Output

id = 5name = cs110owner = 503device =channel =net_path = 192.168.168.102celerra_id = APM000446038450000passphrase = nasadmin

Page 43: Celerra_Replcatr

43 of 172Version 5.6.47Using Celerra Replicator (V1)

important. Be aware when you start to replicate a file system, the facility must create two checkpoints, otherwise replication will not start. For example, if you have 95 checkpoints and want to start a replication, the 96th checkpoint will be created, but replication will fail when the system tries to create the 97th checkpoint because the limit is breached.

◆ Also, when scheduling, be careful not to keep any checkpoints that will surpass the limit otherwise you can not start a replication. In other words, if all checkpoints you specify to keep are created, they must be within the limit.

Action

To create a SnapSure checkpoint of the source file system, use this command syntax:$ fs_ckpt <fs_name> -Create

where:<fs_name> = file system name on which a checkpoint is created

Remote replication example:

To create a checkpoint of the source file system src_ufs1, type:$ fs_ckpt src_ufs1 -Create

Local replication example:

To create a checkpoint of the source file system local_src, type:$ fs_ckpt local_src -Create

Note: The output shown is for the remote replication example.

Page 44: Celerra_Replcatr

Using Celerra Replicator (V1)44 of 172 Version 5.6.47

Output

operation in progress (not interruptible)...id = 88name = src_ufs1acl = 0in_use = Truetype = uxfsworm = offvolume = v243pool = clar_r5_performancemember_of = root_avm_fs_group_10rw_servers= server_2ro_servers=rw_vdms =ro_vdms =ckpts = src_ufs1_ckpt1stor_devs = APM00034000068-001F,APM00034000068-001Edisks = d21,d15 disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14 server=server_2 disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13 server=server_2

id = 90name = src_ufs1_ckpt1acl = 0in_use = Truetype = ckptworm = offvolume = vp246pool = clar_r5_performancemember_of =rw_servers=ro_servers= server_2rw_vdms =ro_vdms =checkpt_of= src_ufs1 Mon Feb 7 06:58:10 EST 2005used = 1%full(mark)= 90%stor_devs = APM00034000068-001F,APM00034000068-001Edisks = d21,d15 disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14 server=server_2 disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13 server=server_2

Page 45: Celerra_Replcatr

45 of 172Version 5.6.47Using Celerra Replicator (V1)

Task 4: Create the destination file system

The destination file system is created as rawfs because replication starts on a rawfs file system. When all steps to set up replication are complete, the file system will be a normal uxfs type. The destination file system must be the same size as the source file system. Managing Celerra Volumes and File Systems Manually and the appropriate man page describe how to create and use a file system.

Task 5: Copy checkpoint to the destination file system

This task copies the entire checkpoint of the source file system created in Task 3: "Create SnapSure checkpoint of source file system" on page 42 to the destination file system. This creates a baseline copy of the source file system on the destination file system. This copy is updated incrementally with changes occurring to the source file system.

Perform this copy task once per file system that is to be replicated. The checkpoint must be copied without converting it to uxfs, by using the convert=no option.

Use of the monitor=off option runs this command as a background process, allowing you to run several copy sessions simultaneously.

Note: If the primary file system extends during the running of the fs_copy command and before replication starts in Task 6: "Begin replication" on page 48, you must extend the destination file system manually to keep file system sizes identical. Use the nas_fs -xtend command.

Step Action

1. Create a destination file system using the samesize= option by typing:$ nas_fs -name dst_ufs1 -type rawfs -create samesize=src_ufs1:cel=cs100 pool=clar_r5_performance

Local replication example:$ nas_fs -name local_dst -type rawfs -create samesize=local_source pool=clar_r5_performance

2. Create a mount point on the destination Data Mover by typing:$ server_mountpoint server_2 -create /dst ufs_1

Local replication example:$ server_mountpoint server_2 -create /local_dst

3. Mount the file system as read-only on the destination Data Mover by typing:$ server_mount server_2 -option ro dst_ufs1 /dst_ufs1

Local replication example:$ server_mount server_2 -option ro local_dst /local_dst

Note: The destination file system can only be mounted on one Data Mover, even though it is read-only.

Page 46: Celerra_Replcatr

Using Celerra Replicator (V1)46 of 172 Version 5.6.47

Action

To copy a checkpoint to the destination file system, use this command syntax:$ fs_copy -start <srcfs> <dstfs>:cel=<cel_name> -option convert=no,monitor=off

where:<srcfs> = source file system checkpoint.<dstfs> = destination file system.<cel_name> = destination Celerra Network Server name.-option convert=[yes|no] = allows the conversion of the <dstfs> to uxfs after the file system copy is executed. If no is specified, when the copy has completed, the <dstfs> remains a rawfs file system type. The default is yes.-option monitor=off = progress of the copy is printed to the screen by default. The off option forces the command to run as a background process.

Remote replication example:

To copy the checkpoint, src_ufs1_ckpt1, to the destination file system without converting it to uxfs, type:$ fs_copy -start src_ufs1_ckpt1 dst_ufs1:cel=cs110 -option convert=no,monitor=off

Local replication example:

To copy the checkpoint, local_src_ckpt1, to the destination file system, local_dst, type:$ fs_copy -start local_src_ckpt1 local_dst -option convert=no, monitor=off

Note: The output shown is for the remote replication example.

Output

operation in progress (not interruptible)...id = 90name = src_ufs1_ckpt1acl = 0in_use = Truetype = ckptworm = offvolume = vp246pool = clar_r5_performancemember_of =rw_servers=ro_servers= server_2rw_vdms =ro_vdms =checkpt_of= src_ufs1 Mon Feb 7 06:58:10 EST 2005used = 1%full(mark)= 90%stor_devs = APM00034000068-001F,APM00034000068-001Edisks = d21,d15 disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14 server=server_2 disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13 server=server_2

Page 47: Celerra_Replcatr

47 of 172Version 5.6.47Using Celerra Replicator (V1)

Verify the copied checkpoint to the destination file system

id = 126name = dst_ufs1acl = 0in_use = Truetype = rawfsworm = offvolume = v272pool = clar_r5_performancemember_of = root_avm_fs_group_3:cs110rw_servers=ro_servers= server_2rw_vdms =ro_vdms =backup_of = src_ufs1 Mon Feb 7 06:58:10 EST 2005stor_devs = APM00044603845-0008,APM00044603845-0007disks = d8,d9 disk=d8 stor_dev=APM00044603845-0008 addr=c0t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c32t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c16t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c48t1l2 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c16t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c48t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c0t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c32t1l1 server=server_2

done

Action

To verify that the fs_copy command completes, type:$ fs_copy -list

Output

Local Source FilesystemsId Source Destination Status %Remaining CommState Local Destination FilesystemsId Source Destination Status %Remaining CommState

Note

The fs_copy session is not listed in the output indicating the copy is complete.

Output

Page 48: Celerra_Replcatr

Using Celerra Replicator (V1)48 of 172 Version 5.6.47

Task 6: Begin replication

When you start replication, the system verifies that primary and secondary Data Movers can communicate with each other. Next, it starts replicating and then begins tracking all changes made to the source file system. You start this process once per file system to be replicated.

Set your replication policies when you establish this replication relationship. "Replication policies" on page 29 describes this feature. If you want to specify a specific interface or IP address for the replication relationship, do so when you start replication. If you specify an interface for the source site, replication uses that interface until it is changed by the user. If you allow the system to select the interface, it can change to keep the replication relationship running. For example, if the network interface currently being used becomes unavailable, the system attempts to select another interface. If one is found, the replication relationship continues to function. The destination interface, regardless of how selected, is unchanged by the system.

Any future changes to this information requires suspending and restarting replication, as detailed in "Suspend a replication relationship" on page 81 and "Restarting a replication relationship" on page 89.

Action

To start replication for the first time, use this command syntax:$ fs_replicate -start <srcfs> <dstfs>:cel=<cel_name>

where:<srcfs> = source file system<dstfs> = destination file system<cel_name> = destination Celerra Network Server

Note: Multiple fs_replicate -start processes must be executed sequentially, not in parallel. Run only one fs_replicate -start command at a time.

Remote replication example:

To start replication for source file system src_ufs1 and a destination file system dst_ufs1, type:$ fs_replicate -start src_ufs1 dst_ufs1:cel=cs110

Local replication example:

To start replication for source file system local_src and a destination file system local_dst, type:$ fs_replicate -start local_src local_dst

Note: The output shown is for the remote replication example.

Page 49: Celerra_Replcatr

49 of 172Version 5.6.47Using Celerra Replicator (V1)

Output

operation in progress (not interruptible)...id = 88name = src_ufs1acl = 0in_use = Truetype = uxfsworm = offvolume = v243pool = clar_r5_performancemember_of = root_avm_fs_group_10rw_servers= server_2ro_servers=rw_vdms =ro_vdms =ckpts = src_ufs1_ckpt1ip_copies = dst_ufs1:cs110stor_devs = APM00034000068-001F,APM00034000068-001Edisks = d21,d15 disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14 server=server_2 disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13 server=server_2

id = 126name = dst_ufs1acl = 0in_use = Truetype = rawfsworm = offvolume = v272pool = clar_r5_performancemember_of = root_avm_fs_group_3:cs110rw_servers=ro_servers= server_2rw_vdms =ro_vdms =backup_of = src_ufs1 Mon Feb 7 06:58:10 EST 2005stor_devs = APM00044603845-0008,APM00044603845-0007disks = d8,d9 disk=d8 stor_dev=APM00044603845-0008 addr=c0t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c32t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c16t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c48t1l2 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c16t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c48t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c0t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c32t1l1 server=server_2done

Page 50: Celerra_Replcatr

Using Celerra Replicator (V1)50 of 172 Version 5.6.47

Task 7: Create a second checkpoint of the source file system

The changes between the two checkpoints are copied to the destination file system in the next task.

Note

The system selects the interface if you do not specify one.

Error messages:

• If you receive an error message stating the interface is not configured or is invalid, the IP addresses for the interface ports are not configured on the destination site. Define these interface ports by running the server_ifconfig command at the destination site.

• If Error 2211: Sec: invalid id specified appears, the local and destination sites have different passphrases. To modify this, follow the procedure in "Change the passphrase between Celerra Network Servers" on page 124.

Action

To create a second checkpoint of the source file system, which is compared to the initial checkpoint, use this command syntax:$ fs_ckpt <fs_name> -Create

where:<fs_name> = file system name for which a checkpoint is created

Remote replication example:

To create a SnapSure checkpoint of source file system src_ufs1, type:$ fs_ckpt src_ufs1 -Create

Local replication example:

To create a SnapSure checkpoint of source file system local_src, type:$ fs_ckpt local_src -Create

Note: The following output shown is for the remote replication example.

Page 51: Celerra_Replcatr

51 of 172Version 5.6.47Using Celerra Replicator (V1)

Output

operation in progress (not interruptible)...id = 88name = src_ufs1acl = 0in_use = Truetype = uxfsworm = offvolume = v243pool = clar_r5_performancemember_of = root_avm_fs_group_10rw_servers= server_2ro_servers=rw_vdms =ro_vdms =ckpts = src_ufs1_ckpt1,src_ufs1_ckpt2ip_copies = dst_ufs1:cs110stor_devs = APM00034000068-001F,APM00034000068-001Edisks = d21,d15 disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14 server=server_2 disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13 server=server_2

id = 97name = src_ufs1_ckpt2acl = 0in_use = Truetype = ckptworm = offvolume = vp246pool = clar_r5_performancemember_of =rw_servers=ro_servers= server_2rw_vdms =ro_vdms =checkpt_of= src_ufs1 Mon Feb 7 07:05:00 EST 2005used = 3%full(mark)= 90%stor_devs = APM00034000068-001F,APM00034000068-001Edisks = d21,d15 disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14 server=server_2 disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13 server=server_2

Page 52: Celerra_Replcatr

Using Celerra Replicator (V1)52 of 172 Version 5.6.47

Task 8: Copy incremental changes

To ensure that the file system type is uxfs when the copy completes, do not use the convert=no option. Use of the monitor=off option runs this command as a background process, which enables you to run several copy sessions simultaneously.

Action

To copy the incremental changes (or the delta set) between the two source file system checkpoints to the destination file system, use this command syntax:$ fs_copy -start <new_check_point> <dstfs>:cel=<cel_name> -fromfs <previous_check_point> -option monitor=off

where:<new_check_point> = last checkpoint taken as described in Task 7: "Create a second checkpoint of the source file system" on page 50.<dstfs> = destination file system.<cel_name> = Celerra Network Server where the destination file system resides.<previous_check_point> = first checkpoint taken.-option monitor=off = progress of the copy is printed to the screen by default. The off option forces the command to run as a background process.

Remote replication example:

The command to use here differs, depending on how you copied your source file system to the destination site. Use the -Force option only if you used a physical transport method (disk or tape).

To copy the incremental changes between checkpoints of source file system src_ufs1, to destination file system dst_ufs1, type:$ fs_copy -start src_ufs1_ckpt2 dst_ufs1:cel=cs110 -fromfs src_ufs1_ckpt1 -option monitor=off

If you used a physical transport to perform the original source file system copy to the destination site, type:$ fs_copy -start src_ufs1_ckpt2 dst_ufs1:cel=cs110 -fromfs src_ufs1_ckpt1 -Force -option monitor=off

Local replication example:

To copy the incremental changes between checkpoints of source file system local_src, type:$ fs_copy -start local_src_ckpt2 local_dst -fromfs local_src_ckpt1 -option monitor=off

Note: The following output shown is for the remote replication example.

Page 53: Celerra_Replcatr

53 of 172Version 5.6.47Using Celerra Replicator (V1)

Output

operation in progress (not interruptible)...id = 97name = src_ufs1_ckpt2acl = 0in_use = Truetype = ckptworm = offvolume = vp246pool = clar_r5_performancemember_of =rw_servers=ro_servers= server_2rw_vdms =ro_vdms =checkpt_of= src_ufs1 Mon Feb 7 07:05:00 EST 2005used = 3%full(mark)= 90%stor_devs = APM00034000068-001F,APM00034000068-001Edisks . = d21,d15 disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14 server=server_2 disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13 server=server_2

id = 126name = dst_ufs1acl = 0in_use = Truetype = rawfsworm = offvolume = v272pool = clar_r5_performancemember_of = root_avm_fs_group_3:cs110rw_servers=ro_servers= server_2rw_vdms =ro_vdms =backup_of = src_ufs1 Mon Feb 7 07:05:00 EST 2005stor_devs = APM00044603845-0008,APM00044603845-0007disks = d8,d9 disk=d8 stor_dev=APM00044603845-0008 addr=c0t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c32t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c16t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c48t1l2 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c16t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c48t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c0t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c32t1l1 server=server_2

done

Page 54: Celerra_Replcatr

Using Celerra Replicator (V1)54 of 172 Version 5.6.47

Verify the copied incremental changes

Task 9: Verify file system conversion

At the destination site, verify file system conversion:

Action

To verify that the fs_copy command completes, type:$ fs_copy -list

Output

Local Source FilesystemsId Source Destination Status %Remaining CommState

Local Destination FilesystemsId Source Destination Status %Remaining CommState

Note

The fs_copy session is not listed in the output which indicates that the copy is complete.

Action

To verify that the file system is converted to a uxfs type file system, type:[destination_site]$ nas_fs -info dst_ufs1

Page 55: Celerra_Replcatr

55 of 172Version 5.6.47Using Celerra Replicator (V1)

Task 10: Check replication status

You can optionally check the status of all replication sessions that are running on a Celerra system and an individual replication session:

◆ "List all replication sessions (optional)" on page 55

◆ "List individual replication session (optional)" on page 57

List all replication sessions (optional)

Output

id = 126name = dst_ufs1acl = 0in_use = Truetype = uxfsworm = offvolume = v272pool = clar_r5_performancemember_of = root_avm_fs_group_3:cs110rw_servers=ro_servers= server_2rw_vdms =ro_vdms =backup_of = src_ufs1 Mon Feb 7 07:05:00 EST 2005stor_devs = APM00044603845-0008,APM00044603845-0007disks = d8,d9 disk=d8 stor_dev=APM00044603845-0008 addr=c0t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c32t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c16t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c48t1l2 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c16t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c48t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c0t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c32t1l1 server=server_2

Action

To list the current active replication sessions, type:$ fs_replicate -list

Page 56: Celerra_Replcatr

Using Celerra Replicator (V1)56 of 172 Version 5.6.47

Output

Source:

Local Source FilesystemsId Source FlowCtrl State Destination FlowCtrl State Network138 src_ufs1 inactive active dst_ufs1:cs110 inactive active alive

Local Destination FilesystemsId Source FlowCtrl State Dest. FlowCtrl State Network

Destination:

Local Source FilesystemsId Source FlowCtrl State Destination FlowCtrl State Network

Local Destination FilesystemsId Source FlowCtrl State Dest. FlowCtrl State Network165 src_ufs1:cs100 inactive active dst_ufs1 inactive active alive

Note

For local replication, the Destination FlowCtrl status and Network status always contain N/A (not applicable).

Page 57: Celerra_Replcatr

57 of 172Version 5.6.47Using Celerra Replicator (V1)

List individual replication session (optional)

Execute this command from either Celerra Network Server for which the system reports the replication status of each current delta set on source and destination sites.

Action

To check the replication status and generate historical data about the replication up to the number of lines specified, use this command syntax:$ fs_replicate -info <fs_name> -verbose <number_of_lines>

where:<fs_name> = name of the file system (in the example it is the source file system).<number_of_lines> = lines to display historical replication data. The maximum number is 128.

Remote replication example:

To check the replication status of the replication relationship, type:$ fs_replicate -info src_ufs1 -verbose 10

Local replication example:

To check the replication status of the replication relationship, type:$ fs_replicate -info local_src -verbose 10

Note: The following output shown is for the remote replication example. "Appendix A: fs_replicate -info output fields" on page 164 gives the output definition description for the fs_replicate -info command.

Page 58: Celerra_Replcatr

Using Celerra Replicator (V1)58 of 172 Version 5.6.47

Output

id = 88name = src_ufs1fs_state = activetype = replicationreplicator_state = activesource_policy = NoPolicyhigh_water_mark = 600time_out = 600current_delta_set = 3current_number_of_blocks = 1flow_control = inactivetotal_savevol_space = 1048576 KBytessavevol_space_available = 917504 KBytes (Before Flow Control)

id = 126name = dst_ufs1:cs110type = playbackplayback_state = activehigh_water_mark = 600time_out = 600current_delta_set = 3flow_control = inactivetotal_savevol_space = 1048576 KBytessavevol_space_available = 786432 KBytes (Before Flow Control)

outstanding delta sets: <None>

communication_state = alivecurrent_transfer_rate = ~ 13312 Kbits/secondavg_transfer_rate = ~ 169984 Kbits/secondsource_ip = 192.168.168.18source_port = 57273destination_ip = 192.168.168.20destination_port = 8888QOS_bandwidth = 0 kbits/sec

| Source | DestinationDelta|Create Time Dur Blocks|Playback Time Dur Blocks DSinGroup-----|-------------- ------ ------|-------------- ------ ------ ---------2 2005-02-08 00:20 11 2005-02-08 00:20 128 2005-02-08 00:20 128 20 2005-02-08 00:10 333 2005-02-08 00:10 333 1

Note

All times are GMT. Block size is 8 KB.

Page 59: Celerra_Replcatr

59 of 172Version 5.6.47Using Celerra Replicator (V1)

Task 11: Create restartable checkpoints

To ensure that a valid checkpoint is available to restart an out-of-sync replication relationship, create two checkpoints of the source file system. These restartable checkpoints, named <source_fs_name>_repl_restart_1 and <source_fs_name>_repl_restart_2, are automatically refreshed by a system CRON job that runs every hour at 25 minutes after the hour. When a replication relationship is out-of-sync, these checkpoints are available for the restart process.

You must include these two checkpoints when determining the maximum number of checkpoints per file system used with replication.

Action

To create checkpoints for use when a replication relationship falls out-of-sync, use this command syntax:$ fs_ckpt <fs_name> -name <name> -Create

where:<fs_name> = file system name on which a checkpoint is created<name> = name of the refresh checkpoint that must follow this naming convention <source_fs_name>_repl_restart_1 and <source_fs_name>_repl_restart_2

Example:

To create the checkpoints of the source file system for use when a replication relationship falls out-of-sync, type:$ fs_ckpt src_ufs1 -name src_ufs1_repl_restart_1 -Create$ fs_ckpt src_ufs1 -name src_usf1_repl_restart_2 -Create

The output only shows the creation of the first checkpoint.

Page 60: Celerra_Replcatr

Using Celerra Replicator (V1)60 of 172 Version 5.6.47

Output

operation in progress (not interruptible)...id = 88name = src_ufs1acl = 0in_use = Truetype = uxfsworm = offvolume = v243pool = clar_r5_performancemember_of = root_avm_fs_group_10rw_servers= server_2ro_servers=rw_vdms =ro_vdms =ckpts = src_ufs1_ckpt1,src_ufs1_ckpt2,src_ufs1_repl_restart_1ip_copies = dst_ufs1:cs110stor_devs = APM00034000068-001F,APM00034000068-001Edisks = d21,d15 disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14 server=server_2 disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13 server=server_2

id = 100name = src_ufs1_repl_restart_1acl = 0in_use = Truetype = ckptworm = offvolume = vp246pool = clar_r5_performancemember_of =rw_servers=ro_servers= server_2rw_vdms =ro_vdms =checkpt_of= src_ufs1 Mon Feb 7 07:14:26 EST 2005used = 4%full(mark)= 90%delta_number= 1stor_devs = APM00034000068-001F,APM00034000068-001Edisks = d21,d15 disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14 server=server_2 disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13 server=server_2

Page 61: Celerra_Replcatr

61 of 172Version 5.6.47Using Celerra Replicator (V1)

Recovering replication data

Occasionally, you might need to recover from a situation in which a replication relationship must be terminated because of a disaster situation. For example, the source site experiences a disaster and becomes unavailable for data processing, or the source file system becomes corrupted and is unrecoverable. In this situation:

1. You can perform a failover from the site where you replicate your data. A checkpoint is created on the destination site which allows the system to track changes that occur after the failover. The failover process makes the destination file system read/write. "Replication failover" on page 62 provides more information.

2. When the source site becomes available again, resynchronize your source and destination sites and restart replication. However, after resynchronization replication is running in the reverse direction. "Resynchronize the source and destination sites" on page 66 provides more information.

3. Use a replication reversal to restore the direction of replication to what it was prior to the failover. When you schedule a replication reversal, the write activity on the destination file system is stopped, and the changes are applied to the source file system before the source site becomes read/write. The reversal process synchronizes the source and destination file systems. "Replication reversal" on page 74 provides more information.

Note: Replication failover, resynchronization, and reversal must be performed sequentially. For example, if you replicate multiple file systems from one Celerra system to another, you must run separate failover commands and wait for one command to complete before running the next command.

Table 7 on page 61 shows the relationships between the source and destination file systems when there is a initiated replication, failover, resynchronization, and reversal.

Table 7 Replication file system relationships (page 1 of 2)

Replication option

Source site Destination site Explanation

Begin replication Source file system is read/write.

Destination file system is read-only.

Normal replication processing establishes source file system as read/write and destination file system as read-only.

Failover Source file system becomes read-only, if source site is still available.

Destination file system becomes read/write.

Changes which file system acts as the source file system and which acts as the destination file system.

Brings the destination file system to read/write to service the I/O in the case of disaster.

Page 62: Celerra_Replcatr

Using Celerra Replicator (V1)62 of 172 Version 5.6.47

Task 1: Replication failover

Failover is the process that changes the destination file system from read-only to read/write and stops the transmission of replicated data. The source file system, if available, becomes read-only. Use the failover operation if the source site is unavailable due to a disaster or if the source site is still available but you want to activate the destination file system as read/write. Perform a failover using the options specified in Table 8 on page 62.

Resynchronize Source file system remains read-only.

Destination file system remains read/write.

Use to reestablish replication after a failover.

Both sites must be available.

Reversal The read-only file system becomes read/write.

The read/write file system becomes read-only.

Changes which file system acts as the source file system and which acts as the destination file system.

Perform a reversal from whichever file system is read/write.

When used after a failover, restores the direction of replication to what it was prior to the failover.

Both sites must be available.

Table 8 Failover options (page 1 of 2)

Failover option

Use if What happensSite that must be available

default The source site is totally corrupt or unavailable.

Plays back all available delta sets at the destination site before failing over.

Destination site

now The source site is totally corrupt or unavailable.

Initiates an immediate failover and no delta sets are played back.

Note: If you perform a failover using this option and delta sets are in the SavVol at the destination site, an incremental resynchronization might not be possible in all cases.

Destination site

Table 7 Replication file system relationships (page 2 of 2)

Replication option

Source site Destination site Explanation

Page 63: Celerra_Replcatr

63 of 172Version 5.6.47Using Celerra Replicator (V1)

Using the failover command when the source site is unavailable results in data loss because delta sets cannot be transferred from the source site to the destination site. Using the default failover option reduces the amount of data loss, by replaying any available delta sets (pending data) at the destination before initiating failover.

Failover processing creates a file system checkpoint to use to resynchronize the replication relationship. You must include this checkpoint when determining the maximum number of checkpoints per file system used with replication.

When a failover completes, replication is stopped and the destination file system becomes read/write. Replication is no longer running because the source site is usually unavailable when a failover is initiated. If the source site becomes available, reestablish replication in the opposite direction (from the destination site to the source site) by resynchronizing the source and destination file systems.

For replication failover:

◆ "Verify status of destination file system" on page 64

◆ "Initiate replication failover" on page 64

◆ "Verify file system is read/write" on page 66

sync The source site is still available.

Fails over without any data loss by making the source file system read-only, the destination file system read/write, and creating a restart checkpoint of the destination file system.

Synchronized failover takes longer to invoke and cannot be performed if the source site is unavailable. It is more suited to a maintenance-related failover as part of a failover plan.

Note: The sync option is not used in a disaster situation because both sites must be available.

Source and destination

Table 8 Failover options (page 2 of 2)

Failover option

Use if What happensSite that must be available

Page 64: Celerra_Replcatr

Using Celerra Replicator (V1)64 of 172 Version 5.6.47

Verify status of destination file system

Initiate replication failover

Action

To check the status of the destination file system and verify if it is mounted as read-only, use this command syntax:

$ server_mount <movername>

where:<movername> = name of the Data Mover on which the file system is mounted

Example:

To verify the status of the destination file systems on server_2, type:

$ server_mount server_2

Note: The ro in the output indicates a read-only file system.

Output

server_2 :root_fs_2 on / uxfs,perm,rwfs1 on /fs1 uxfs,perm,rwckpt1 on /ckpt1 ckpt,perm,rofsk on /fsk ckpt,perm,roroot_fs_common on /.etc_common uxfs,perm,rodst_ufsl on /dst_ufsl uxfs,perm,ro

Action

To initiate a failover from the destination site, use this command syntax:[destination_site]$ fs_replicate -failover <srcfs>:cel=<cel_name> <dstfs> -option now|sync

where:<srcfs> = source file system name<cel_name> = Celerra Network Server name of the source site<dstfs> = destination file system name

Example:

To fail over the source file system to the destination file system, type:$ fs_replicate -failover src_ufs1:cel=cs100 dst_ufs1

Page 65: Celerra_Replcatr

65 of 172Version 5.6.47Using Celerra Replicator (V1)

Output

operation in progress (not interruptible)...id = 126name = dst_ufs1acl = 0in_use = Truetype = uxfsworm = offvolume = v272pool = clar_r5_performancemember_of = root_avm_fs_group_3rw_servers= server_2ro_servers=rw_vdms =ro_vdms =ckpts = root_restart_ckpt_88_2stor_devs = APM00044603845-0008,APM00044603845-0007disks = d8,d9 disk=d8 stor_dev=APM00044603845-0008 addr=c0t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c32t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c16t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c48t1l2 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c16t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c48t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c0t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c32t1l1 server=server_2

id = 88name = src_ufs1acl = 0in_use = Truetype = uxfsworm = offvolume = v243pool = clar_r5_performancemember_of = root_avm_fs_group_10:cs100rw_servers= server_2ro_servers=rw_vdms =ro_vdms =ckpts = src_ufs1_repl_restart_1:cs100,src_ufs1_repl_restart_2:cs100stor_devs = APM00034000068-001F,APM00034000068-001Edisks = d21,d15 disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14 server=server_2 disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13 server=server_2done

Page 66: Celerra_Replcatr

Using Celerra Replicator (V1)66 of 172 Version 5.6.47

Verify file system is read/write

Task 2: Resynchronize the source and destination sites

After the failover command is run, a checkpoint is created on the destination site and the destination file system becomes read/write. New writes are then allowed on the destination file system.

When replication is resynchronized, default values are used for the replication policies. For example, high water mark and timeout are set to 600. You can specify new policies when you restart replication using the any of these options:

◆ autofullcopy={yes}

◆ to=<timeout>

◆ dto=<destination timeout>

◆ hwm=<high_water_mark>

◆ dhwm=<destination high_water_mark>

◆ qos=<qos>

◆ autoro={yes|no}

◆ autofreeze={yes|no}

Note: If you need to increase your file system’s size and plan to resynchronize your source and destination sites after a failover, you must complete the resynchronization (fs_replicate -resync) before increasing the size of your destination file system.

To resynchronize the source and destination sites:

◆ "Verify file system is read-only" on page 67

◆ "Resynchronize the source and destination file systems and restart replication" on page 68

Action

To verify that the file system is mounted as read/write and is accessible to the network clients, type:$ server_mount server_2

Output

server_mount server_2server_2 :root_fs_2 on / uxfs,perm,rwroot_fs_common on /.etc_common uxfs,perm,rodst_ufs1 on /dst_ufs1 uxfs,perm,rw

Page 67: Celerra_Replcatr

67 of 172Version 5.6.47Using Celerra Replicator (V1)

Verify file system is read-only

Step Action Result

1. Before resynchronizing the file systems, verify whether the file system on the original source site is mounted as read-only by typing:$ server_mount server_2

Note: If the source file system is not mounted read-only, the following message appears:Error 4124: <file_system> : is not mounted ro

server_2 :root_fs_2 on / uxfs,perm,rwroot_fs_common on /.etc_common uxfs,perm,rosrc_ufs1 on /src_ufs1 uxfs,perm,ro

2. If the file system (src_ufs1) in the example is read/write, change it to read-only by typing:$ server_mount server_2 -option ro src_ufs1

server_2 : done

Page 68: Celerra_Replcatr

Using Celerra Replicator (V1)68 of 172 Version 5.6.47

Resynchronize the source and destination file systems and restart replication

Action

To attempt to resynchronize the source and destination file systems and restart replication from the destination site, use this command syntax:[destination_site]$ fs_replicate -resync <dstfs>[:cel=<cel_name>] <srcfs> -option autofullcopy=yes

where:<dstfs> = current read-only file system name<cel_name> = name of Celerra Network Server of the original source site<srcfs> = current read/write file system name-option autofullcopy=yes = executes a full copy of the file system if an incremental resynchronization does not complete.

Example:

To resynchronize the file systems and resume replication, type:$ fs_replicate -resync src_ufs1:cel=cs100 dst_ufs1

For the system to automatically perform a full copy of the source file system if the incremental resynchronization fails, type:$ fs_replicate -resync src_ufs1:cel=cs100 dst_ufs1 -option autofullcopy=yes

The full copy of the file system using autofullcopy=yes can be time-consuming. Consider when you want to run this command.

Note: If a disaster occurs during the transfer, some delta sets might become lost. As a result, the replication process will not be able to completely replicate the source file system on the destination site.

Output: Convert to rawfs

operation in progress (not interruptible)...Converting filesystem typeid = 88name = src_ufs1acl = 0in_use = Truetype = rawfsworm = offvolume = v243pool = clar_r5_performancemember_of = root_avm_fs_group_10rw_servers=ro_servers= server_2rw_vdms =ro_vdms =ckpts = src_ufs1_repl_restart_2,src_ufs1_repl_restart_1stor_devs = APM00034000068-001F,APM00034000068-001Edisks = d21,d15 disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14 server=server_2 disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13 server=server_2

Page 69: Celerra_Replcatr

69 of 172Version 5.6.47Using Celerra Replicator (V1)

Output: Start copy

Starting baseline copy...operation in progress (not interruptible)...id = 133name = root_restart_ckpt_88_2acl = 0in_use = Truetype = ckptworm = offvolume = vp284pool = clar_r5_performancemember_of =rw_servers=ro_servers= server_2rw_vdms =ro_vdms =checkpt_of= dst_ufs1 Mon Feb 7 07:25:11 EST 2005used = 1%full(mark)= 90%delta_number= 4stor_devs = APM00044603845-0008,APM00044603845-0007disks = d8,d9 disk=d8 stor_dev=APM00044603845-0008 addr=c0t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c32t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c16t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c48t1l2 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c16t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c48t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c0t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c32t1l1 server=server_2

id = 88name = src_ufs1acl = 0in_use = Truetype = rawfsworm = offvolume = v243pool = clar_r5_performancemember_of = root_avm_fs_group_10:cs100rw_servers=ro_servers= server_2rw_vdms =ro_vdms =backup_of = dst_ufs1 Mon Feb 7 07:25:11 EST 2005ckpts = src_ufs1_repl_restart_2:cs100,src_ufs1_repl_restart_1:cs100stor_devs = APM00034000068-001F,APM00034000068-001Edisks = d21,d15 disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14 server=server_2 disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13 server=server_2

IP Copy remaining (%) 100..Done.

done

Page 70: Celerra_Replcatr

Using Celerra Replicator (V1)70 of 172 Version 5.6.47

Output: Start copy

operation in progress (not interruptible)...id = 88name = src_ufs1acl = 0in_use = Truetype = rawfsworm = offvolume = v243pool = clar_r5_performancemember_of = root_avm_fs_group_10rw_servers=ro_servers= server_2rw_vdms =ro_vdms =backup_of = dst_ufs1:cs110 Mon Feb 7 07:25:11 EST 2005ckpts = src_ufs1_repl_restart_2,src_ufs1_repl_restart_1stor_devs = APM00034000068-001F,APM00034000068-001Edisks = d21,d15 disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14 server=server_2 disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13 server=server_2done

Page 71: Celerra_Replcatr

71 of 172Version 5.6.47Using Celerra Replicator (V1)

Output: Starting replication

Starting replication...operation in progress (not interruptible)...id = 126name = dst_ufs1acl = 0in_use = Truetype = uxfsworm = offvolume = v272pool = clar_r5_performancemember_of = root_avm_fs_group_3rw_servers= server_2ro_servers=rw_vdms =ro_vdms =ckpts = root_restart_ckpt_88_2ip_copies = src_ufs1:cs100stor_devs = APM00044603845-0008,APM00044603845-0007disks = d8,d9 disk=d8 stor_dev=APM00044603845-0008 addr=c0t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c32t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c16t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c48t1l2 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c16t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c48t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c0t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c32t1l1 server=server_2

id = 88name = src_ufs1acl = 0in_use = Truetype = rawfsworm = offvolume = v243pool = clar_r5_performancemember_of = root_avm_fs_group_10:cs100rw_servers=ro_servers= server_2rw_vdms =ro_vdms =backup_of = dst_ufs1 Mon Feb 7 07:25:11 EST 2005ckpts = src_ufs1_repl_restart_2:cs100,src_ufs1_repl_restart_1:cs100stor_devs = APM00034000068-001F,APM00034000068-001Edisks = d21,d15 disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14 server=server_2 disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13 server=server_2done

Page 72: Celerra_Replcatr

Using Celerra Replicator (V1)72 of 172 Version 5.6.47

Output: Generating a new checkpoint

Generating new checkpoint...operation in progress (not interruptible)...id = 126name = dst_ufs1acl = 0in_use = Truetype = uxfsworm = offvolume = v272pool = clar_r5_performancemember_of = root_avm_fs_group_3rw_servers= server_2ro_servers=rw_vdms =ro_vdms =ckpts = root_restart_ckpt_88_2,root_new_ckpt_dst_ufs1ip_copies = src_ufs1:cs100stor_devs = APM00044603845-0008,APM00044603845-0007disks = d8,d9 disk=d8 stor_dev=APM00044603845-0008 addr=c0t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c32t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c16t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c48t1l2 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c16t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c48t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c0t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c32t1l1 server=server_2

id = 140name = root_new_ckpt_dst_ufs1acl = 0in_use = Truetype = ckptworm = offvolume = vp284pool = clar_r5_performancemember_of =rw_servers=ro_servers= server_2rw_vdms =ro_vdms =checkpt_of= dst_ufs1 Mon Feb 7 07:34:31 EST 2005used = 3%full(mark)= 90%stor_devs = APM00044603845-0008,APM00044603845-0007disks = d8,d9 disk=d8 stor_dev=APM00044603845-0008 addr=c0t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c32t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c16t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c48t1l2 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c16t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c48t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c0t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c32t1l1 server=server_2

Page 73: Celerra_Replcatr

73 of 172Version 5.6.47Using Celerra Replicator (V1)

Output: Starting differential copy

Starting diff copy...

operation in progress (not interruptible)...id = 140name = root_new_ckpt_dst_ufs1acl = 0in_use = Truetype = ckptworm = offvolume = vp284pool = clar_r5_performancemember_of =rw_servers=ro_servers= server_2rw_vdms =ro_vdms =checkpt_of= dst_ufs1 Mon Feb 7 07:34:31 EST 2005used = 3%full(mark)= 90%stor_devs = APM00044603845-0008,APM00044603845-0007disks = d8,d9 disk=d8 stor_dev=APM00044603845-0008 addr=c0t1l2

server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c32t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c16t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c48t1l2 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c16t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c48t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c0t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c32t1l1 server=server_2

id = 88name = src_ufs1acl = 0in_use = Truetype = rawfsworm = offvolume = v243pool = clar_r5_performancemember_of = root_avm_fs_group_10:cs100rw_servers=ro_servers= server_2rw_vdms =ro_vdms =backup_of = dst_ufs1 Mon Feb 7 07:34:31 EST 2005ckpts = src_ufs1_repl_restart_2:cs100,src_ufs1_repl_restart_1:cs100stor_devs = APM00034000068-001F,APM00034000068-001Edisks = d21,d15 disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14 server=server_2 disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13 server=server_2

IP Copy remaining (%) 100..Done.

done

Page 74: Celerra_Replcatr

Using Celerra Replicator (V1)74 of 172 Version 5.6.47

Task 3: Replication reversal

A replication reversal is a scheduled change in the direction of replication and requires both sites to be available. If you successfully resumed replication using the -resync option, replication is now running in the reverse direction—from the destination site to the source site. You can now execute a reversal to change the direction of replication. This reversal returns the original source file system to read/write and the original destination file system to read-only.

The write activity on the destination file system is stopped and any changes are applied to the source file system before the source site becomes read/write. The reversal process keeps the source and destination file systems synchronized. A reversal is usually performed after a failover and resynchronization, but it can be used at any time to reverse the replication direction.

Note: This can only be done when both sites are operational.

For replication reversal:

◆ "Verify direction of replication process" on page 75

◆ "Reverse the replication" on page 76

Output: Deleting checkpoints that are used to restart replication

Deleting root_restart_ckpt_88_2...id = 133name = root_restart_ckpt_88_2acl = 0in_use = Falsetype = ckptworm = offvolume =rw_servers=ro_servers=rw_vdms =ro_vdms =Deleting root_new_ckpt_dst_ufs1...id = 140name = root_new_ckpt_dst_ufs1acl = 0in_use = Falsetype = ckptworm = offvolume =rw_servers=ro_servers=rw_vdms =ro_vdms =Operation completedone

Page 75: Celerra_Replcatr

75 of 172Version 5.6.47Using Celerra Replicator (V1)

Verify direction of replication process

Before you reverse the replication direction, verify which file system is read/write and which is read-only by using the server_mount server_x command. There will be a rw or ro entry to indicate the file system’s status. You then verify the direction.

Action

To verify the direction of your replication relationship, type:$ fs_replicate -list

Note: Use this command on the site where the file system is read/write. In this example, the destination file system is read/write because of the failover.

Output

fs_replicate -listLocal Source FilesystemsId Source FlowCtrl State Destination FlowCtrl State Network7 dst_ufs1 inactive active src_ufs1:cs100 inactive active alive

Local Destination FilesystemsId Source FlowCtrl State Destination FlowCtrl State Network

Note

Notice the file system named dst_ufs1 is acting as the source file system (read/write) and the file system named src_ufs1 is functioning as the destination file system (read-only).

Page 76: Celerra_Replcatr

Using Celerra Replicator (V1)76 of 172 Version 5.6.47

Reverse the replication

Action

To initiate a reversal from the original destination site, use this command syntax:[original_destination_site]$ fs_replicate -reverse <dstfs>:cel=<cel_name> <srcfs>

where:<dstfs> = current read-only file system name. This file system will become the read/write file system.<cel_name> = name of Celerra Network Server at the original source site.<srcfs> = current read/write file system name. This file system will become the read-only file system.

Example:

To reverse the direction of the replication process and return the src_ufs1 file system to read/write and the dst_ufs1 file system to read-only, type:$ fs_replicate -reverse src_ufs1:cel=cs100 dst_ufs1

Page 77: Celerra_Replcatr

77 of 172Version 5.6.47Using Celerra Replicator (V1)

Output

operation in progress (not interruptible)...id = 126name = dst_ufs1acl = 0in_use = Truetype = uxfsworm = offvolume = v272pool = clar_r5_performancemember_of = root_avm_fs_group_3rw_servers= server_2ro_servers=rw_vdms =ro_vdms =ip_copies = src_ufs1:cs100stor_devs = APM00044603845-0008,APM00044603845-0007disks = d8,d9 disk=d8 stor_dev=APM00044603845-0008 addr=c0t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c32t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c16t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c48t1l2 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c16t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c48t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c0t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c32t1l1 server=server_2

id = 88name = src_ufs1acl = 0in_use = Truetype = uxfsworm = offvolume = v243pool = clar_r5_performancemember_of = root_avm_fs_group_10:cs100rw_servers=ro_servers= server_2rw_vdms =ro_vdms =backup_of = dst_ufs1 Mon Feb 7 07:34:31 EST 2005ckpts = src_ufs1_repl_restart_2:cs100,src_ufs1_repl_restart_1:cs100stor_devs = APM00034000068-001F,APM00034000068-001Edisks = d21,d15 disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14 server=server_2 disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13 server=server_2done

operation in progress (not interruptible)...done

Page 78: Celerra_Replcatr

Using Celerra Replicator (V1)78 of 172 Version 5.6.47

After a failover or reversal

After you complete a failover or reversal, to enable users to access the file system that has newly acquired read/write access permission, you must configure Celerra system appropriately for NFS. In many cases, you can use the same configuration you used for the source file system which might include:

◆ File system exports

◆ I18N mode (Unicode or ASCII) configuration

◆ Network interfaces

Replicating EMC Celerra CIFS Environments (V1) describes all CIFS replication issues.

Note

• After this command completes, the original source file system is read/write and the original destination file system is read-only. Replication is now running in the direction it was before the failover. You can verify this by using the server_mount server_x command.

• When replication is reversed, default values are used for the replication policies. For example, high water mark and timeout are set to 600. You can specify new policies when you restart replication using -option <options>.

Page 79: Celerra_Replcatr

79 of 172Version 5.6.47Using Celerra Replicator (V1)

Abort Celerra Replicator

Abort Celerra Replicator when you no longer want to replicate the file system or when your source and destination file systems are not synchronized and you want to end replication.

After Celerra Replicator successfully aborts, the destination file system is a read-only file system at the destination site. The source file system is a read/write file system at the source site. When a session has been aborted it cannot be restarted using the-restart option.

Note: Aborting replication does not delete the underlying file systems.

Note: Multiple fs_replicate -abort processes are executed sequentially, not in parallel. Only run one fs_replicate -abort command at a time.

Action

To abort replication on source and destination file systems simultaneously, use this command syntax from the source site:$ fs_replicate -abort <srcfs>,<dstfs>:cel=<cel_name>

where:<srcfs> = name of the source file system<dstfs> = name of the destination file system <cel_name> = name of the remote Celerra Network Server

Example:

To stop replication for the replication relationship, type:[source_site]$ fs_replicate -abort src_ufs1,dst_ufs1:cel=cs110

Page 80: Celerra_Replcatr

Using Celerra Replicator (V1)80 of 172 Version 5.6.47

Output

operation in progress (not interruptible)...id = 88name = src_ufs1acl = 0in_use = Truetype = uxfsworm = offvolume = v243pool = clar_r5_performancemember_of = root_avm_fs_group_10rw_servers= server_2ro_servers=rw_vdms =ro_vdms =ckpts = src_ufs1_repl_restart_2,src_ufs1_repl_restart_1stor_devs = APM00034000068-001F,APM00034000068-001Edisks = d21,d15 disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14 server=server_2 disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13 server=server_2

id = 126name = dst_ufs1acl = 0in_use = Truetype = uxfsworm = offvolume = v272pool = clar_r5_performancemember_of = root_avm_fs_group_3:cs110rw_servers=ro_servers= server_2rw_vdms =ro_vdms =stor_devs = APM00044603845-0008,APM00044603845-0007disks = d8,d9 disk=d8 stor_dev=APM00044603845-0008 addr=c0t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c32t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c16t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c48t1l2 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c16t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c48t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c0t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c32t1l1 server=server_2done

Page 81: Celerra_Replcatr

81 of 172Version 5.6.47Using Celerra Replicator (V1)

Suspend a replication relationship

Suspend is an option that allows you to temporarily stop an active replication relationship and leave replication in a condition that allows it to be restarted.

Action

To suspend replication, use this command syntax:[source_site]$ fs_replicate -suspend <srcfs> <dstfs>:cel=<cel_name>

where:<scrfs> = name of the source file system<dstfs> = name of the destination file system<cel_name> = name of the destination Celerra Network Server

Example:

To suspend a replication relationship, type:

$ fs_replicate -suspend src_ufs1 dst_ufs1:cel=cs110

Page 82: Celerra_Replcatr

Using Celerra Replicator (V1)82 of 172 Version 5.6.47

Suspend output: Creating a new delta set

operation in progress (not interruptible)...operation in progress (not interruptible)...id = 88name = src_ufs1fs_state = activetype = replicationreplicator_state = activesource_policy = NoPolicyhigh_water_mark = 0time_out = 0current_delta_set = 5current_number_of_blocks = 0flow_control = inactivetotal_savevol_space = 1048576 KBytessavevol_space_available = 917504 KBytes (Before Flow Control)

id = 126name = dst_ufs1:cs110type = playbackplayback_state = activehigh_water_mark = 600time_out = 600current_delta_set = 4flow_control = inactivetotal_savevol_space = 1048576 KBytessavevol_space_available = 917504 KBytes (Before Flow Control)

outstanding delta sets:Delta Source_create_time Blocks----- ------------------ ------4 2005-02-09 06:54:26 1

communication_state = alivecurrent_transfer_rate = ~ 13312 Kbits/secondavg_transfer_rate = ~ 13312 Kbits/secondsource_ip = 192.168.168.18source_port = 62817destination_ip = 192.168.168.20destination_port = 8888QOS_bandwidth = 0 kbits/sec

Note: All times are in GMT. Block size is 8 KBytes.

done

Page 83: Celerra_Replcatr

83 of 172Version 5.6.47Using Celerra Replicator (V1)

Suspend output: Creating a baseline checkpoint

Generating new checkpoint...operation in progress (not interruptible)...id = 88name = src_ufs1acl = 0in_use = Truetype = uxfsworm = offvolume = v243pool = clar_r5_performancemember_of = root_avm_fs_group_10rw_servers= server_2ro_servers=rw_vdms =ro_vdms =ckpts = src_ufs1_repl_restart_1,src_ufs1_ckpt1,src_ufs1_repl_restart_2,root_susp end_ckpt_126_5ip_copies = dst_ufs1:cs110stor_devs = APM00034000068-001F,APM00034000068-001Edisks = d21,d15 disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14 server=server_2 disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13 server=server_2

id = 125name = root_suspend_ckpt_126_5acl = 0in_use = Truetype = ckptworm = offvolume = vp246pool = clar_r5_performancemember_of =rw_servers=ro_servers= server_2rw_vdms =ro_vdms =checkpt_of= src_ufs1 Tue Feb 8 13:51:01 EST 2005used = 6%full(mark)= 90%delta_number= 5stor_devs = APM00034000068-001F,APM00034000068-001Edisks = d21,d15 disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14 server=server_2 disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13 server=server_2

Page 84: Celerra_Replcatr

Using Celerra Replicator (V1)84 of 172 Version 5.6.47

Suspend output: Creating another delta set

operation in progress (not interruptible)...id = 88name = src_ufs1fs_state = activetype = replicationreplicator_state = activesource_policy = NoPolicyhigh_water_mark = 0time_out = 0current_delta_set = 6current_number_of_blocks = 0flow_control = inactivetotal_savevol_space = 1048576 KBytessavevol_space_available = 917504 KBytes (Before Flow Control)

id = 126name = dst_ufs1:cs110type = playbackplayback_state = activehigh_water_mark = 600time_out = 600current_delta_set = 4flow_control = inactivetotal_savevol_space = 1048576 KBytessavevol_space_available = 786432 KBytes (Before Flow Control)

outstanding delta sets:Delta Source_create_time Blocks----- ------------------ ------5 2005-02-09 06:54:53 14 2005-02-09 06:54:26 1

communication_state = alive current_transfer_rate = ~ 13312 Kbits/secondavg_transfer_rate = ~ 13312 Kbits/secondsource_ip = 192.168.168.18source_port = 62817destination_ip = 192.168.168.20destination_port = 8888QOS_bandwidth = 0 kbits/sec

Note: All times are in GMT. Block size is 8 KBytes.

done

Page 85: Celerra_Replcatr

85 of 172Version 5.6.47Using Celerra Replicator (V1)

Suspend output: Playing back delta set

operation in progress (not interruptible)...id = 88name = src_ufs1:cs100fs_state = activetype = replicationreplicator_state = activesource_policy = NoPolicyhigh_water_mark = 0time_out = 0current_delta_set = 6current_number_of_blocks = 0flow_control = inactivetotal_savevol_space = 1048576 KBytessavevol_space_available = 917504 KBytes (Before Flow Control)

id = 126name = dst_ufs1type = playbackplayback_state = activehigh_water_mark = 0time_out = 10current_delta_set = 6flow_control = inactivetotal_savevol_space = 1048576 KBytessavevol_space_available = 786432 KBytes (Before Flow Control)

outstanding delta sets: <None>

communication_state = alivecurrent_transfer_rate = ~ 13312 Kbits/secondavg_transfer_rate = ~ 13312 Kbits/secondsource_ip = 192.168.168.18source_port = 62817destination_ip = 192.168.168.20destination_port = 8888QOS_bandwidth = 0 kbits/sec

Note: All times are in GMT. Block size is 8 KBytes.

done

Page 86: Celerra_Replcatr

Using Celerra Replicator (V1)86 of 172 Version 5.6.47

Suspend output: Waiting for synchronization

operation in progress (not interruptible)...id = 88name = src_ufs1acl = 0in_use = Truetype = uxfsworm = offvolume = v243pool = clar_r5_performancemember_of = root_avm_fs_group_10rw_servers= server_2ro_servers=rw_vdms =ro_vdms =ckpts = src_ufs1_repl_restart_1,src_ufs1_ckpt1,src_ufs1_repl_restart_2,root_susp end_ckpt_126_5ip_copies = dst_ufs1:cs110stor_devs = APM00034000068-001F,APM00034000068-001Edisks = d21,d15 disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14 server=server_2 disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13 server=server_2done

Page 87: Celerra_Replcatr

87 of 172Version 5.6.47Using Celerra Replicator (V1)

Suspend output: Convert destination file system to rawfs — ready for restart

name = dst_ufs1acl = 0in_use = Truetype = uxfsworm = offvolume = v272pool = clar_r5_performancemember_of = root_avm_fs_group_3rw_servers=ro_servers= server_2rw_vdms =ro_vdms =stor_devs = APM00044603845-0008,APM00044603845-0007disks = d8,d9 disk=d8 stor_dev=APM00044603845-0008 addr=c0t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c32t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c16t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c48t1l2 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c16t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c48t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c0t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c32t1l1 server=server_2doneConverting filesystem typeid = 126name = dst_ufs1acl = 0in_use = Truetype = rawfsworm = offvolume = v272pool = clar_r5_performancemember_of = root_avm_fs_group_3rw_servers=ro_servers= server_2rw_vdms =ro_vdms =stor_devs = APM00044603845-0008,APM00044603845-0007disks = d8,d9 disk=d8 stor_dev=APM00044603845-0008 addr=c0t1l2

server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c32t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c16t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c48t1l2 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c16t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c48t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c0t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c32t1l1 server=server_2Operation completedone

Page 88: Celerra_Replcatr

Using Celerra Replicator (V1)88 of 172 Version 5.6.47

Verify the suspended replication relationship

Action

To verify that the replication relationship no longer exists, type either of the following two commands. The -list option no longer displays the replication pair:$ fs_replicate -list$ fs_replicate -info src_ufs1

Output

Local Source FilesystemsId Source FlowCtrl State Destination FlowCtrl State Network

Local Destination FilesystemsId Source FlowCtrl State Dest. FlowCtrl State Network

The replication session is no longer listed.

Error 2242: src_ufs1 : replication/playback is not set up

Page 89: Celerra_Replcatr

89 of 172Version 5.6.47Using Celerra Replicator (V1)

Restarting a replication relationship

To restart a replication relationship:

◆ "Verify that the replication relationship is not synchronized" on page 90

◆ "Restart replication relationship" on page 91

Page 90: Celerra_Replcatr

Using Celerra Replicator (V1)90 of 172 Version 5.6.47

Verify that the replication relationship is not synchronized

Action

To verify the replication relationship’s status, use this command syntax:$ fs_replicate -info <srcfs>

where:<srcfs> = name of the source file system

Example:

To verify the replication relationship, type:$ fs_replicate -info src_ufs1 -verbose 10

Output

id = 88name = src_ufs1fs_state = activetype = replicationreplicator_state = inactivesource_policy = NoPolicyhigh_water_mark = 600time_out = 600current_delta_set = 0current_number_of_blocks = 0flow_control = activetotal_savevol_space = 1048576 KBytessavevol_space_available = 0 KBytes (Before Flow Control)

id = 126name = dst_ufs1:cs110type = playbackplayback_state = activehigh_water_mark = 600time_out = 600current_delta_set = 146flow_control = inactivetotal_savevol_space = 1048576 KBytessavevol_space_available = 917504 KBytes (Before Flow Control)

outstanding delta sets: <None>

communication_state = downcurrent_transfer_rate = ~ 0 Kbits/secondavg_transfer_rate = ~ 0 Kbits/secondsource_ip = 0.0.0.0source_port = 0destination_ip = 192.168.168.20destination_port = 8888QOS_bandwidth = 0 kbits/sec

| Source | DestinationDelta|Create Time Dur Blocks|Playback Time Dur Blocks DSinGroup-----|-------------- ------ ------|-------------- ------ ------ ---------

Note: All times are in GMT. Block size is 8 KBytes.

Page 91: Celerra_Replcatr

91 of 172Version 5.6.47Using Celerra Replicator (V1)

Restart replication relationship

Action

To restart a replication relationship, use this command syntax from the source site:$ fs_replicate -restart <srcfs> <dstfs>:cel=<cel_name>

where:<scrfs> = name of the source file system<dstfs> = name of the destination file system<cel_name> = name of the destination Celerra Network Server

Example:

To restart a replication relationship, type:$ fs_replicate -restart src_ufs1 dst_ufs1:cel=cs110

Note: The following output shown is for restarting an out-of-synchronization replication.

Page 92: Celerra_Replcatr

Using Celerra Replicator (V1)92 of 172 Version 5.6.47

Restart output: Converting file system type

operation in progress (not interruptible)...operation in progress (not interruptible)...id = 88name = src_ufs1acl = 0in_use = Truetype = uxfsworm = offvolume = v243pool = clar_r5_performancemember_of = root_avm_fs_group_10rw_servers= server_2ro_servers=rw_vdms =ro_vdms =ckpts = src_ufs1_repl_restart_2,src_ufs1_repl_restart_1ip_copies = dst_ufs1:cs110stor_devs = APM00034000068-001F,APM00034000068-001Edisks = d21,d15 disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14 server=server_2 disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13 server=server_2done

operation in progress (not interruptible)...id = 126name = dst_ufs1acl = 0in_use = Truetype = uxfsworm = offvolume = v272pool = clar_r5_performancemember_of = root_avm_fs_group_3rw_servers=ro_servers= server_2rw_vdms =ro_vdms =stor_devs = APM00044603845-0008,APM00044603845-0007disks = d8,d9disk=d8 stor_dev=APM00044603845-0008 addr=c0t1l2server=server_2disk=d8 stor_dev=APM00044603845-0008 addr=c32t1l2 server=server_2disk=d8 stor_dev=APM00044603845-0008 addr=c16t1l2 server=server_2disk=d8 stor_dev=APM00044603845-0008 addr=c48t1l2 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c16t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c48t1l1 server=server_2disk=d9 stor_dev=APM00044603845-0007 addr=c0t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c32t1l1 server=server_2done

Page 93: Celerra_Replcatr

93 of 172Version 5.6.47Using Celerra Replicator (V1)

Converting filesystem typeid = 126name = dst_ufs1acl = 0in_use = Truetype = rawfsworm = offvolume = v272pool = clar_r5_performancemember_of = root_avm_fs_group_3rw_servers=ro_servers= server_2rw_vdms =ro_vdms =

Restart output: Converting file system type

stor_devs = APM00044603845-0008,APM00044603845-0007disks = d8,d9 disk=d8 stor_dev=APM00044603845-0008 addr=c0t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c32t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c16t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c48t1l2

server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c16t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c48t1l1

server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c0t1l1

server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c32t1l1 server=server_2

Restart output: Converting file system type

Page 94: Celerra_Replcatr

Using Celerra Replicator (V1)94 of 172 Version 5.6.47

Restart output: Starting replication

Starting replication...operation in progress (not interruptible)...id = 88name = src_ufs1acl = 0in_use = Truetype = uxfsworm = offvolume = v243pool = clar_r5_performancemember_of = root_avm_fs_group_10rw_servers= server_2ro_servers=rw_vdms =ro_vdms =ckpts = src_ufs1_repl_restart_2,src_ufs1_repl_restart_1ip_copies = dst_ufs1:cs110stor_devs = APM00034000068-001F,APM00034000068-001Edisks = d21,d15 disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14 server=server_2 disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13 server=server_2

id = 126name = dst_ufs1acl = 0in_use = Truetype = rawfsworm = offvolume = v272pool = clar_r5_performancemember_of = root_avm_fs_group_3:cs110rw_servers=ro_servers= server_2rw_vdms =ro_vdms =backup_of = src_ufs1 Tue Feb 8 08:49:46 EST 2005stor_devs = APM00044603845-0008,APM00044603845-0007disks = d8,d9 disk=d8 stor_dev=APM00044603845-0008 addr=c0t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c32t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c16t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c48t1l2 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c16t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c48t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c0t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c32t1l1 server=server_2done

Page 95: Celerra_Replcatr

95 of 172Version 5.6.47Using Celerra Replicator (V1)

Restart output: Create checkpoint

Generating new checkpoint...operation in progress (not interruptible)...id = 88name = src_ufs1acl = 0in_use = Truetype = uxfsworm = offvolume = v243pool = clar_r5_performancemember_of = root_avm_fs_group_10rw_servers= server_2ro_servers=rw_vdms =ro_vdms =ckpts = src_ufs1_repl_restart_2,src_ufs1_repl_restart_1,root_new_ckpt_src_ufs1ip_copies = dst_ufs1:cs110stor_devs = APM00034000068-001F,APM00034000068-001Edisks = d21,d15 disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14 server=server_2 disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13 server=server_2

id = 115name = root_new_ckpt_src_ufs1acl = 0in_use = Truetype = ckptworm = offvolume = vp246pool = clar_r5_performancemember_of =rw_servers=ro_servers= server_2rw_vdms =ro_vdms =checkpt_of= src_ufs1 Tue Feb 8 08:50:13 EST 2005used = 4%full(mark)= 90%stor_devs = APM00034000068-001F,APM00034000068-001Edisks = d21,d15 disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14 server=server_2 disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13 server=server_2

Page 96: Celerra_Replcatr

Using Celerra Replicator (V1)96 of 172 Version 5.6.47

Restart output: Start differential copy

Starting diff copy...operation in progress (not interruptible)...id = 115name = root_new_ckpt_src_ufs1acl = 0in_use = Truetype = ckptworm = offvolume = vp246pool = clar_r5_performancemember_of =rw_servers=ro_servers= server_2rw_vdms =ro_vdms =checkpt_of= src_ufs1 Tue Feb 8 08:50:13 EST 2005used = 4%full(mark)= 90%stor_devs = APM00034000068-001F,APM00034000068-001Edisks = d21,d15 disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14 server=server_2 disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13 server=server_2

id = 126name = dst_ufs1acl = 0in_use = Truetype = rawfsworm = offvolume = v272pool = clar_r5_performancemember_of = root_avm_fs_group_3:cs110rw_servers=ro_servers= server_2rw_vdms =ro_vdms =backup_of = src_ufs1 Tue Feb 8 08:50:13 EST 2005stor_devs = APM00044603845-0008,APM00044603845-0007disks = d8,d9 disk=d8 stor_dev=APM00044603845-0008 addr=c0t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c32t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c16t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c48t1l2 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c16t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c48t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c0t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c32t1l1 server=server_2

IP Copy remaining (%) 100..Done.

done

Page 97: Celerra_Replcatr

97 of 172Version 5.6.47Using Celerra Replicator (V1)

Restart output: Delete restart checkpoints

Deleting root_new_ckpt_src_ufs1...id = 115name = root_new_ckpt_src_ufs1acl = 0in_use = Falsetype = ckptworm = offvolume =rw_servers=ro_servers=rw_vdms =ro_vdms =Operation completedone

Page 98: Celerra_Replcatr

Using Celerra Replicator (V1)98 of 172 Version 5.6.47

Extending the size of a file system

File systems can be extended automatically and manually. Follow these procedures to extend file system size:

◆ "Extend file system size automatically" on page 98

◆ "Extend file system size" on page 101

Note: You cannot extend file systems using non-sliced volumes if replication is running. When using Automatic File System Extension, the slice option must be enabled for an Automatic Volume Management (AVM) pool when replication is running. Managing EMC Celerra Volumes and File Systems with Automatic Volume Management describes how to configure slices in detail.

Extend file system size automatically

File systems in a replication relationship can be extended automatically. For replication, the Automatic File System Extension policy can be set on the source file system only. When the policy is set, the destination file system is extended, and then the source file system, just as with extending file system manually. This feature is used on file systems created using AVM only. Managing EMC Celerra Volumes and File Systems with Automatic Volume Management details Automatic File System Extension.

Automatic File System Extension provides the following options in the CLI and in Celerra Manager (they are called out here as they appear in Celerra Manager):

◆ Auto Extend enabled: To select the function.

◆ Virtual Provisioning™ enabled: To allocate storage capacity based on anticipated need, but dedicate resources only as needed. When virtual provisioning is enabled, the maximum file system size or real file system size, whichever is larger is reported to NFS and CIFS clients through Celerra Manager or CLI.

◆ High water mark: To specify the threshold, ranging from 50 to 90 percent of the file system, which triggers file system extension.

◆ Maximum capacity (MB): The maximum size the file system can be extended.

Note: The virtual size of a source but not destination file system is visible from the NFS/CIFS client.

Page 99: Celerra_Replcatr

99 of 172Version 5.6.47Using Celerra Replicator (V1)

Action

To automatically extend the size of a file system (setting a high water mark, maximum size, and virtual provisioning), use this command syntax:$ nas_fs -modify <fs_name> -auto_extend yes -hwm <50-99>% -max_size <integer>[T|G|M] -vp <yes|no>

where:<fs_name> = name of the file system.<50-99> = percentage of the file system that is full and in need of being reached before the file system is automatically extended.<integer> = file system’s maximum size (entered in TB, GB, or MB).-vp = used with a specified Maximum Capacity value to report the anticipated or actual file system size. When turned on, the virtual size is reported to clients.

Example:

To automatically extend the source file system src_ufs1 with a high water mark of 50% and a minimum size of 70 MB with virtual provisioning enabled, type:$ nas_fs -modify src_ufs1 -auto_extend yes -hwm 50% -max_size 70M -vp yes

Output

id = 2707name = pfs001acl = 0in_use = Truetype = uxfsworm = offvolume = v7283pool = clarata_archivemember_of = root_avm_fs_group_10rw_servers= 123secnfsro_servers=rw_vdms =ro_vdms =auto_ext = hwm=50%,max_size=16777216M,virtual_provision=yesckpts = pfs001_ckpt60,pfs001_ckpt61,pfs001_ckpt62,pfs001_ckpt63,pfs001_ckpt6stor_devs = APM00043200225-0029,APM00043200225-002C,APM00043200225-0027,APM00043disks = d35,d21,d34,d20,d32,d16,d30,d14,d33,d17,d31,d15,d38,d41,d36,d22,d29,d13,d39,d42,d28,d12,d37,d23disk=d35 stor_dev=APM00043200225-0029 addr=c16t2l9 server=123secnfs[nasadmin@lnsgc123 nasadmin]$

Page 100: Celerra_Replcatr

Using Celerra Replicator (V1)100 of 172 Version 5.6.47

Recover from Automatic File System Extension failure

Automatic File System Extension employs an internal script that checks for adequate space on the source and destination sites. If replication is thwarted because extension of either file system fails due to insufficient space, an error message displays the cause. For example:

Nov 21 10:37:53 2005 CFS:3:101 fs auto extension failed: no space available to extend src1

In the great majority of Automatic File System Extension failures, both file systems fail to extend. By consulting the sys_log, you can determine whether only the source file system extension failed if the log contains a line that directs you to use the src_only option.

Whether one or both file system extensions fail, find the file systems’ sizes and compare their block counts and volume sizes by using the nas_fs -size command. The number displayed is a second way to determine which file system extension failed. This method allows you to know how much space to reserve when you manually extend the file systems. Perform the following steps to recover from a failure to automatically extend the file system.

Step Action

1. Display the sys_log by typing:$ cd /nas/log/$ more sys_log

2. Find the size of the source and destination file system, and compare the values using this command syntax:$ nas_fs -size <fs_name>

The following example shows sample output:total = 100837 avail = 99176 used = 1660 ( 1% ) (sizes in MB) ( blockcount = 209715200 )volume: total = 102400 (sizes in MB) ( blockcount = 209715200 )

Note: Because total, available, and used values are generated from the operating system and not updated until the destination file system is refreshed, that data will differ from total volume data derived by the Control Station. Block # counts agree on the same file system, however, and can be used to accurately compare totals on source and destination file systems.

3. To extend the source file system, type the following command from the source site. For size, type the difference between source and destination file systems determined in step 2:$ nas_fs -xtend <fs_name> size=<integer>[T|G|M] -option src_only

Note: Because replication always extends the destination first, the source size will never be larger than the destination size. If you specify the wrong size when you run this command, the command will report the error and you must rerun the command with the correct size.

Page 101: Celerra_Replcatr

101 of 172Version 5.6.47Using Celerra Replicator (V1)

Extend file system size

To extend file system size:

◆ "Extend file system size manually" on page 101

◆ "Extend a file system after replication has been suspended" on page 103

◆ "Extend a file system after replication failover" on page 104

◆ "Start replication when the source file system is inactive" on page 104

Extend file system size manually

File systems can be manually extended while replication is running. If only the destination file system extends, you can extend the source file system manually. The procedure to do this (and extend both file systems) is the same for manual and Automatic File System Extension. Extending the source file system extends the destination file system by default. This maintains the same file system sizes at both sites. "Recover from Automatic File System Extension failure" on page 100 provides more information.

Before extending the source file system, verify that the destination Celerra system contains enough unused space. Use the nas_fs -size command to determine the current size of the file system.

Action

To extend the source file system, use this command syntax:[source_site]$ nas_fs -xtend <fs_name> size=<integer>[T|G|M] pool=<pool>

where:<fs_name> = name of the source file system<integer> = file system’s size in terabytes, gigabytes, or megabytes<pool> = assigns a rule set for the file system

Example:

To extend the source file system src_ufs1 by 1024 megabytes, type:$ nas_fs -xtend src_ufs1 size=1024M pool=clar_r5_performance

Page 102: Celerra_Replcatr

Using Celerra Replicator (V1)102 of 172 Version 5.6.47

Output

id = 88name = src_ufs1acl = 0in_use = Truetype = uxfsworm = offvolume = v243pool = clar_r5_performancemember_of = root_avm_fs_group_10rw_servers= server_2ro_servers=rw_vdms =ro_vdms =ckpts = src_ufs1_repl_restart_2,src_ufs1_repl_restart_1ip_copies = dst_ufs1:cs110stor_devs = APM00034000068-001F,APM00034000068-001Edisks = d21,d15 disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14 server=server_2 disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13 server=server_2

Source file system size:

total = 3025 avail = 2010 used = 1014 ( 33% ) (sizes in MB) ( blockcount = 6291456 )volume: total = 3072 (sizes in MB) ( blockcount = 6291456 )

Destination file system size:

total = 3025 avail = 2010 used = 1014 ( 33% ) (sizes in MB) ( blockcount = 6291456 )volume: total = 3072 (sizes in MB) ( blockcount = 6291456 )

Note: Both file systems have the same block # count (6291456).

The nas_fs -xtend command extends the destination file system and then the source file system. If either file system fails to extend, an error message displays the cause. If neither file system extends, the following sample output appears:

Error 5008: Remote command failed:remote celerra = cs0remote exit status = 5remote error = 0remote message = CLSTD : volume(s) are not available

Note: Step 3 of "Recover from Automatic File System Extension failure" on page 100 explains how to extend both file systems.

If only the destination file system extends, the following output appears:

Error 5008: pfsAvolume(s) are not availablePFS extension failed. Please extend PFS with “nas_fs -xtend... -option src_only”

Note: Go to step 3 of the "Recover from Automatic File System Extension failure" on page 100 to extend only the source file system.

Page 103: Celerra_Replcatr

103 of 172Version 5.6.47Using Celerra Replicator (V1)

Extend a file system after replication has been suspended

If the source file system is extended automatically or manually after replication has been suspended, perform the following steps to ensure that the source and destination file systems are the same size before restarting replication.

If you do not perform these steps before restarting replication, the restart will fail with an error because the source file system and destination file system are no longer the same size (that is, the block counts differ). In this case, you must perform these steps to recover from the error.

Step Action

1. Verify that the destination file system type is set to rawfs. If it is set to uxfs, convert the destination file system from uxfs to rawfs by using this command syntax:$ nas_fs -Type rawfs <dstfs> -Force

where:<dstfs> = name of the destination file system

Example:

To verify that the file system is set to rawfs, type:$ nas_fs -Type rawfs dst_ufs1 -Force

Note: A read-only file system must be set to rawfs prior to extending a file system or restarting a replication.

2. Extend the destination file system manually using the same size as the source file system by using this command syntax:$ nas_fs -xtend <dst_fs> size=<integer>[T|G|M] -option <options>

where:<dstfs> = name of the destination file system <integer> = size of the secondary file system in terabytes, gigabytes, or megabytes<option> = any comma-separated options, such as slice={y|n}, which specifies whether the disk volumes used by the file system might be shared with other file systems using a slice

Example:

To extend the destination file system dst_ufs1 by 2 MB, using the slice option, to match the source file system extension to the same size, type:$ nas_fs -xtend dst_ufs1 size=2M slice=y

3. Restart the replication relationship by using this command syntax from the source site:$ fs_replicate -restart <srcfs> <dstfs>:cel=<cel_name>

where:<scrfs> = name of the source file system<dstfs> = name of the destination file system<cel_name> = name of the destination Celerra Network Server

Example:

To restart a replication relationship, type:$ fs_replicate -restart src_ufs1 dst_ufs1:cel=cs110

Page 104: Celerra_Replcatr

Using Celerra Replicator (V1)104 of 172 Version 5.6.47

Extend a file system after replication failover

After a replication failover ends, you must reconfigure the Automatic File System Extension policy if you want to continue using the function on the destination file system as your PFS. This is because the Celerra Network Server database, which stores the automatic extension configuration, is not copied to the destination file system during replication. As a result, the destination file system cannot be extended automatically.

The procedure on page 98 provides an example, and Managing EMC Celerra Volumes and File Systems with Automatic Volume Management offers detailed instructions to configure this function’s policy.

If the source file system is still active and the original destination file system is extended, and you want to resynchronize both file systems, follow this procedure:

Start replication when the source file system is inactive

If the source file system is inactive, you have to consider other options. Consult "Abort Celerra Replicator" on page 79 for guidance on whether to attempt an incremental resynchronization of the source and destination file systems or start replication from the beginning.

Step Action

1. Verify that the source file system is mounted read-only. If it is not, do so now using this command syntax:$ server_mount <servername> -option ro <srcfs>/<srcfs_mountpoint>

2. Convert the source file system from uxfs to rawfs using this command syntax:$ nas_fs -Type rawfs <srcfs> -Force

3. Manually extend the source file system and match the extended destination file system size using this command syntax:$ nas_fs -xtend <srcfs> size=<bytes>

4. Convert the source file system back to uxfs using this command syntax:$ nas_fs -Type uxfs <srcfs> -Force

5. Resynchronize the destination and original source file systems using this command syntax:$ fs_replicate -resync <srcfs>:cel=<cel_name> <dstfs>

6. Reverse the replication and return to the original configuration using this command syntax:$ fs_replicate -reverse <dstfs> <dstfs>:cel=<cel_name> <srcfs>

Page 105: Celerra_Replcatr

105 of 172Version 5.6.47Using Celerra Replicator (V1)

Resetting replication policy

Replication policies are established when you initially set up the replication relationship. Use the fs_replicate command to change the policies for each replication session.

To reset replication policy:

◆ "High water mark and time-out policies" on page 105

◆ "Modify replication policy" on page 106

◆ "Change flow-control policies" on page 107

◆ "Set bandwidth size" on page 109

◆ "Set policies using parameters" on page 110

High water mark and time-out policies

High water mark and time-out policies prevent the replication service from creating delta sets faster than it can copy them to the destination file system.

To avoid file systems dropping out-of-sync, reset these source file system policies to higher values than the destination file system policies. The following example resets the high water mark replication policy to:

◆ Trigger the replication service to create delta sets every 300 MB of change for the source file system.

◆ Replay these delta sets to the destination file system every 300 MB of change.

Table 9 on page 105 explains the fs_replicate command options to change the policies.

Note: Multiple fs_replicate -refresh processes must be run sequentially, not concurrently. Run only one fs_replicate -refresh command at a time.

Table 9 Options to set high water mark and time-out policies

Option Behavior

-modify To specify values for the source and destination sites, include both file systems in the command syntax. The values are effective the next time a trigger for these policies is reached. For example, if policies are changed from 600 and 300 for high water mark and time-out interval, respectively, the next time replication reaches 600, the trigger is changed to 300. If you set source high water mark and time-out interval values without specifying values for the destination, the source values are applied to the destination site.

-refresh A refresh replication initiates playback of outstanding data already on the destination site and then creates a delta set of modified data on the source site. This has the same effect as reaching the next high water mark or time-out interval.

Page 106: Celerra_Replcatr

Using Celerra Replicator (V1)106 of 172 Version 5.6.47

Modify replication policy

Action

To modify the replication policy for a source and destination file system, use this command syntax:$ fs_replicate -modify <srcfs> -option hwm=<high_water_mark>, to=<timeout>,dhwm=<high_water_mark>,dto=<timeout>

where:<srcfs> = name of the source file system-option hwm=<high_water_mark> = high water mark policy in megabytes-option dto=<timeout> = time-out policy in seconds

Example:

To reset the high water mark for source and destination file systems, type: $ fs_replicate -modify src_ufs1,dst_ufs1:cel=cs110 -option hwm=300,dhwm=300

Output

operation in progress (not interruptible)...id = 88name = src_ufs1fs_state = activetype = replicationreplicator_state = activesource_policy = NoPolicyhigh_water_mark = 600 (Pending: 300)time_out = 600current_delta_set = 11current_number_of_blocks = 1flow_control = inactivetotal_savevol_space = 1048576 KBytessavevol_space_available = 917504 KBytes (Before Flow Control)

id = 126name = dst_ufs1:cs110type = playbackplayback_state = activehigh_water_mark = 600 (Pending: 300)time_out = 600current_delta_set = 11flow_control = inactivetotal_savevol_space = 1048576 KBytessavevol_space_available = 917504 KBytes (Before Flow Control)

outstanding delta sets: <None>

communication_state = alivecurrent_transfer_rate = ~ 13312 Kbits/secondavg_transfer_rate = ~ 13312 Kbits/secondsource_ip = 172.24.168.123source_port = 62815destination_ip = 192.168.168.20destination_port = 8888QOS_bandwidth = 0 kbits/sec

Note: All times are in GMT. Block size is 8 KBytes.

done

Page 107: Celerra_Replcatr

107 of 172Version 5.6.47Using Celerra Replicator (V1)

Change flow-control policies

When the source SavVol is full and the Data Mover can no longer track changes to the source file system, use one of these two options to keep the replication session active:

◆ Set a policy to freeze the file system, which temporarily halts all I/Os to the source file system until sufficient space is available on the source SavVol.

◆ Set a policy that temporarily halts all writes to the source file system and makes this file system read-only. This condition persists until sufficient space is available on the source SavVol.

Consider the following when establishing these policies:

◆ These values take effect when the next trigger for creating a delta set is reached.

◆ If replication is currently in one of these flow-control states, the new setting takes effect after exiting the read-only or freeze situation.

◆ The autofreeze and autoro options can only be set on the source file system.

◆ Flow-control policies, by default, are not enabled, which can result in replication becoming inactive. If that happens:

• Try to restart your replication relationship using the procedure described in "Restarting a replication relationship" on page 89. If that is impossible, abort (described in "Abort Celerra Replicator" on page 79) and restart replication (described in "Initiating replication" on page 39).

Note

Any changes to the time-out interval or high water mark occur when the next trigger point is reached so the Pending entry shown above is removed when the policy value changes.

Action

To change flow-control policies for a file system, use this command syntax:$ fs_replicate -modify <fs_name>:cel=<cel_name> -option <options>

where:<fs_name> = name of the source file system.<cel_name> = name of the remote Celerra Network Server.<options> = flow-control setting for the source file system. To freeze all I/O to the source file system use autofreeze=yes. To allow users to continue read-only access to the source file system use autoro=yes.

Example:

To freeze all I/O to the source file system, src_ufs1, specify the option autofreeze=yes. Type:$ fs_replicate -modify src_ufs1 -option autofreeze=yes

Page 108: Celerra_Replcatr

Using Celerra Replicator (V1)108 of 172 Version 5.6.47

Output

operation in progress (not interruptible)...id = 88name = src_ufs1fs_state = activetype = replicationreplicator_state = activesource_policy = NoPolicy (Pending: Freeze)high_water_mark = 600time_out = 600current_delta_set = 30current_number_of_blocks = 0flow_control = inactivetotal_savevol_space = 1048576 KBytessavevol_space_available = 917504 KBytes (Before Flow Control)

id = 126name = dst_ufs1:cs110type = playbackplayback_state = activehigh_water_mark = 600time_out = 600current_delta_set = 29flow_control = inactivetotal_savevol_space = 1048576 KBytessavevol_space_available = 917504 KBytes (Before Flow Control)

outstanding delta sets:Delta Source_create_time Blocks----- ------------------ ------29 2005-02-09 11:57:08 1

communication_state = alive current_transfer_rate = ~ 13312 Kbits/secondavg_transfer_rate = ~ 13312 Kbits/secondsource_ip = 192.168.168.18source_port = 62819destination_ip = 192.168.168.20destination_port = 8888QOS_bandwidth = 0 kbits/sec

Note: All times are in GMT. Block size is 8 KBytes.

done

Note

• Determine when the file system is in a read-only or freeze situation by using the fs_replicate -info command and checking the fs_state field.

• When the file system is in a read-only situation, the fs_state field is romounted and when no I/O is allowed, frozen.

Page 109: Celerra_Replcatr

109 of 172Version 5.6.47Using Celerra Replicator (V1)

Set bandwidth size

Setting the bandwidth size, limits the total bandwidth for this replication session. The default value is maximum available network bandwidth 0.

Action

To specify the maximum bandwidth used for a replication session, use this command syntax:$ fs_replicate -modify <fs_name>:cel=<cel_name> -option qos=8000

where:<fs_name> = name of the source file system<cel_name> = name of the remote Celerra Network Server<options> = qos = bandwidth typed in kilobytes per second

Example:

To set the maximum bandwidth for this replication session, type:

$ fs_replicate -modify src_ufs1 -option qos=8000

Note: This setting takes effect the next time data is sent across the IP network.

Page 110: Celerra_Replcatr

Using Celerra Replicator (V1)110 of 172 Version 5.6.47

Set policies using parameters

Use the fs_replicate command to set flow-control policies and bandwidth size for each replication relationship rather than for all replication sessions on a Data Mover. If you choose to set these policies for all sessions on a Data Mover, change the parameter VRPL read-only and VRPL freeze for flow-control policies.

You can view and dynamically modify parameter values using the server_param command or Celerra Manager graphical user interface. This technical module describes only the command-line procedures. Celerra Manager online help details how to use the graphical user interface to modify parameter values.

Output

operation in progress (not interruptible)...id = 88name = src_ufs1fs_state = activetype = replicationreplicator_state = activesource_policy = NoPolicyhigh_water_mark = 600time_out = 600current_delta_set = 30current_number_of_blocks = 1flow_control = inactivetotal_savevol_space = 1048576 KBytessavevol_space_available = 917504 KBytes (Before Flow Control)

id = 126name = dst_ufs1:cs110type = playbackplayback_state = activehigh_water_mark = 600time_out = 600current_delta_set = 30flow_control = inactivetotal_savevol_space = 1048576 KBytessavevol_space_available = 917504 KBytes (Before Flow Control)

outstanding delta sets: <None>

communication_state = alivecurrent_transfer_rate = ~ 13312 Kbits/secondavg_transfer_rate = ~ 13312 Kbits/secondsource_ip = 192.168.168.18source_port = 62819destination_ip = 192.168.168.20destination_port = 8888QOS_bandwidth = 8000 kbits/sec

Note: All times are in GMT. Block size is 8 KBytes.

done

Page 111: Celerra_Replcatr

111 of 172Version 5.6.47Using Celerra Replicator (V1)

Reverse the direction of a replication relationship

Use a replication reversal to change the direction of replication. You might want to change the direction of replication to perform maintenance on the source site or to do testing on the destination site.

Action

To change the direction of replication, use this command syntax:[read/write side]$ fs_replicate -reverse <dstfs>:cel=<cel_name> <srcfs>

where:<dstfs> = name of the file system currently read-only<cel_name> = name of Celerra Network Server where the destination file system currently resides<srcfs> = name of the file system currently read/write

Example:

For the current read/write file system, src_ufs1, to become the read-only file system, type:$ fs_replicate -reverse dst_ufs1:cel=cs110 src_ufs

Page 112: Celerra_Replcatr

Using Celerra Replicator (V1)112 of 172 Version 5.6.47

Output

operation in progress (not interruptible)...id = 88name = src_ufs1acl = 0in_use = Truetype = uxfsworm = offvolume = v243pool = clar_r5_performancemember_of = root_avm_fs_group_10rw_servers= server_2ro_servers=rw_vdms =ro_vdms =ckpts = src_ufs1_ckpt1,src_ufs1_repl_restart_2,src_ufs1_repl_restart_1ip_copies = dst_ufs1:cs110stor_devs = APM00034000068-001F,APM00034000068-001Edisks = d21,d15 disk=d21 stor_dev=APM00034000068-001F addr=c16t1l14 server=server_2 disk=d21 stor_dev=APM00034000068-001F addr=c0t1l14 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c0t1l13 server=server_2 disk=d15 stor_dev=APM00034000068-001E addr=c16t1l13 server=server_2

id = 126name = dst_ufs1acl = 0in_use = Truetype = uxfsworm = offvolume = v272pool = clar_r5_performancemember_of = root_avm_fs_group_3:cs110rw_servers=ro_servers= server_2rw_vdms =ro_vdms =backup_of = src_ufs1 Tue Feb 8 13:53:59 EST 2005stor_devs = APM00044603845-0008,APM00044603845-0007disks = d8,d9 disk=d8 stor_dev=APM00044603845-0008 addr=c0t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c32t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c16t1l2 server=server_2 disk=d8 stor_dev=APM00044603845-0008 addr=c48t1l2 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c16t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c48t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c0t1l1 server=server_2 disk=d9 stor_dev=APM00044603845-0007 addr=c32t1l1 server=server_2done

Page 113: Celerra_Replcatr

113 of 172Version 5.6.47Using Celerra Replicator (V1)

Verify the reverse direction of replication relationship

Note

• When this command completes, the current read/write file system (src_ufs1) becomes read-only and the current read-only file system (dst_ufs1) becomes read/write.

• If you tried to run this command from the incorrect side (read-only), this error message appears:Error 2247: this command must be issued on the current source site:cs100

Action

To verify if the direction of the replication reversed, type:$ fs_replicate -list

Output

Local Source FilesystemsId Source FlowCtrl State Destination FlowCtrl State Network

Local Destination FilesystemsId Source FlowCtrl State Dest. FlowCtrl State Network135 dst_ufs1:cs110 inactive active src_ufs1 inactive active alive

Page 114: Celerra_Replcatr

Using Celerra Replicator (V1)114 of 172 Version 5.6.47

Monitor replication

Table 10 on page 114 shows the commands to use to monitor different aspects of replication. You can also monitor replication using Celerra Manager, which is described in Celerra Manager online help.

Table 10 Ways to monitor replication

Monitor Description Command

Data Movers Returns information about available memory and the CPU idle percentage.

server_sysstat <movername>

File systems Reports the amount of used and available disk space for a file system and reports the amount of a file system’s total capacity that is used.

server_df <movername>

Replication Shows information for all replication sessions, as described in "List all replication sessions (optional)" on page 55.

fs_replicate -list

Replication Shows information for an individual replication session, as detailed in "List individual replication session (optional)" on page 57.

fs_replicate -info <fs_name>

Page 115: Celerra_Replcatr

115 of 172Version 5.6.47Using Celerra Replicator (V1)

Checking playback service and outstanding delta sets

The -info option of fs_replicate lets you verify that the playback service is running and which delta sets have not replayed to the destination file system. The following tasks show how to determine that the playback service is running and to verify the outstanding delta sets:

1. "Determine playback service status" on page 115

2. "Playback delta set" on page 118

3. "Verify delta set" on page 119

Task 1: Determine playback service status

Action

To determine the playback service is running, use this command syntax:$ fs_replicate -info <fs_name> -verbose <number_of_lines>

where:<fs_name> = file system name<number_of_lines> = number of lines to display in the output

Example:

To determine the playback service is running, type:$ fs_replicate -info src -verbose 20

Page 116: Celerra_Replcatr

Using Celerra Replicator (V1)116 of 172 Version 5.6.47

Output

$id = 18name = srcfs_state = activetype = replicationreplicator_state = activesource_policy = NoPolicyhigh_water_mark = 60000time_out = 3600current_delta_set = 92current_number_of_blocks = 1flow_control = inactivetotal_savevol_space = 1048576 KBytessavevol_space_available = 393216 KBytes (Before Flow Control)

id = 30name = dest:eng168102type = playbackplayback_state = activehigh_water_mark = 300time_out = 600current_delta_set = 87flow_control = inactivetotal_savevol_space = 1048576 KBytessavevol_space_available = 393216 KBytes (Before Flow Control)

id = 30name = dest:eng168102type = playbackplayback_state = activehigh_water_mark = 300time_out = 600current_delta_set = 87flow_control = inactivetotal_savevol_space = 1048576 KBytessavevol_space_available = 393216 KBytes (Before Flow Control)

outstanding delta sets:Delta Source_create_time Blocks----- ------------------ ------90 06/04 10:08:54 189 06/04 09:58:54 188 06/04 09:58:43 187 06/04 09:48:43 1445791 06/04 10:10:52 1

communication_state = alivecurrent_transfer_rate = ~ 13312 Kbits/secondavg_transfer_rate = ~ 18140.4 Kbits/secondsource_ip = 10.168.0.11source_port = 59068destination_ip = 10.168.0.180destination_port = 8888QOS_bandwidth = 0 kbits/sec

Page 117: Celerra_Replcatr

117 of 172Version 5.6.47Using Celerra Replicator (V1)

| Source | DestinationDelta|Create Time Dur Blocks|Playback Time Dur Blocks DSinGroup-----|-------------- ------ ------|-------------- ------ ------ ---------91 06/04 10:10:52 0 190 06/04 10:08:54 0 189 06/04 09:58:54 0 188 06/04 09:58:43 0 187 06/04 09:48:43 9 1445786 06/04 09:39:25 0 1 06/04 09:44:35 0 1 185 06/04 09:29:25 0 1 06/04 09:34:35 0 1 184 06/04 09:19:25 0 1 06/04 09:24:35 0 1 183 06/04 09:09:25 0 1 06/04 09:14:35 0 1 182 06/04 08:59:25 0 1 06/04 09:04:34 0 1 181 06/04 08:49:25 0 1 06/04 08:54:34 0 1 180 06/04 08:39:25 0 1 06/04 08:44:34 0 1 179 06/04 08:29:25 0 1 06/04 08:34:34 0 1 178 06/04 08:19:25 0 1 06/04 08:24:34 0 1 177 06/04 08:09:25 0 1 06/04 08:14:34 0 1 176 06/04 07:59:25 0 1 06/04 08:04:34 0 1 175 06/04 07:49:25 0 1 06/04 07:54:34 0 1 174 06/04 07:39:25 0 1 06/04 07:44:34 0 1 173 06/04 07:29:25 0 1 06/04 07:34:34 0 1 172 06/04 07:19:25 0 1 06/04 07:24:34 0 1 1Note: All times are in GMT. Block size is 8 KBytes.

Output

Page 118: Celerra_Replcatr

Using Celerra Replicator (V1)118 of 172 Version 5.6.47

Task 2: Playback delta set

Action

To play back all delta sets up to a specified delta-set number, use this command syntax:

$ fs_replicate -refresh dest -option playuntildelta=91

Note: In this example, the system plays back all delta sets up to 91. Any delta sets greater than that number will not replay.

Output

operation in progress (not interruptible)...id = 18name = src:eng16853fs_state = activetype = replicationreplicator_state = activesource_policy = NoPolicyhigh_water_mark = 60000time_out = 3600current_delta_set = 92current_number_of_blocks = 1flow_control = inactivetotal_savevol_space = 1048576 KBytessavevol_space_available = 393216 KBytes (Before Flow Control)

id = 30name = desttype = playbackplayback_state = activehigh_water_mark = 300time_out = 600current_delta_set = 92flow_control = inactivetotal_savevol_space = 1048576 KBytessavevol_space_available = 393216 KBytes (Before Flow Control)

outstanding delta sets: <None>communication_state = alivecurrent_transfer_rate = ~ 13312 Kbits/secondavg_transfer_rate = ~ 18140.4 Kbits/secondsource_ip = 10.168.0.11source_port = 59068destination_ip = 10.168.0.180destination_port = 8888QOS_bandwidth = 0 kbits/sec

Note: All times are in GMT. Block size is 8 KBytes.

done

Page 119: Celerra_Replcatr

119 of 172Version 5.6.47Using Celerra Replicator (V1)

Task 3: Verify delta set

Action

To verify that the specified delta set was replayed, type:

$ fs_replicate -info src -verbose 20

Output

id = 18name = srcfs_state = activetype = replicationreplicator_state = activesource_policy = NoPolicyhigh_water_mark = 60000time_out = 3600current_delta_set = 92current_number_of_blocks = 1flow_control = inactivetotal_savevol_space = 1048576 KBytessavevol_space_available = 393216 KBytes (Before Flow Control)

id = 30name = dest:eng168102type = playbackplayback_state = activehigh_water_mark = 300time_out = 600current_delta_set = 92flow_control = inactivetotal_savevol_space = 1048576 KBytessavevol_space_available = 393216 KBytes (Before Flow Control)

outstanding delta sets: <None>

communication_state = alivecurrent_transfer_rate = ~ 13312 Kbits/secondavg_transfer_rate = ~ 18140.4 Kbits/secondsource_ip = 10.168.0.11source_port = 59068destination_ip = 10.168.0.180destination_port = 8888QOS_bandwidth = 0 kbits/sec

| Source | DestinationDelta|Create Time Dur Blocks|Playback Time Dur Blocks DSinGroup-----|-------------- ------ ------|-------------- ------ ------ ---------91 06/04 10:10:52 0 1 06/04 10:58:42 2 1 590 06/04 10:08:54 0 1 06/04 10:48:42 2 1 589 06/04 09:58:54 0 1 06/04 10:38:42 2 1 588 06/04 09:58:43 0 1 06/04 10:28:42 2 1 587 06/04 09:48:43 9 14457 06/04 10:18:42 2 14457 586 06/04 09:39:25 0 1 06/04 09:44:35 0 1 185 06/04 09:29:25 0 1 06/04 09:34:35 0 1 184 06/04 09:19:25 0 1 06/04 09:24:35 0 1 183 06/04 09:09:25 0 1 06/04 09:14:35 0 1 182 06/04 08:59:25 0 1 06/04 09:04:34 0 1 181 06/04 08:49:25 0 1 06/04 08:54:34 0 1 180 06/04 08:39:25 0 1 06/04 08:44:34 0 1 1

Page 120: Celerra_Replcatr

Using Celerra Replicator (V1)120 of 172 Version 5.6.47

79 06/04 08:29:25 0 1 06/04 08:34:34 0 1 178 06/04 08:19:25 0 1 06/04 08:24:34 0 1 177 06/04 08:09:25 0 1 06/04 08:14:34 0 1 176 06/04 07:59:25 0 1 06/04 08:04:34 0 1 175 06/04 07:49:25 0 1 06/04 07:54:34 0 1 174 06/04 07:39:25 0 1 06/04 07:44:34 0 1 173 06/04 07:29:25 0 1 06/04 07:34:34 0 1 172 06/04 07:19:25 0 1 06/04 07:24:34 0 1 1Note: All times are in GMT. Block size is 8 KBytes.

Output

Page 121: Celerra_Replcatr

121 of 172Version 5.6.47Using Celerra Replicator (V1)

Events for Celerra Replicator

Celerra Replicator provides the following file system events for use in SNMP traps and email. System event traps and email notifications are configured by the user. Configuring EMC Celerra Events and Notifications covers file system events.

Table 11 on page 121 explains Volume Replication (VRPL) facility events (Facility ID : 77).

Table 11 Celerra Replicator events (page 1 of 2)

Event ID Event Description

0 Replication ok Not in use.

1 Replication on Source Filesystem Inactive

Replication service is inactive.

2 Resync asked by (previous) Destination Filesystem

Request for a replication relationship resynchronization after a failover.

3 Source Filesystem Switch Delta on HWM

Not in use.

4 Destination Filesystem in Error Playback service is inactive.

5 IP Rep Svc failed - Transport IPRepSender stopped.

6 IP Rep Svc NetWork or Receiver Down Source or destination IP network is down.

7 IP Rep Svc Network or Receiver Up Source or destination IP network is up.

8 Rep Svc Source Filesystem Flow on Hold

No source SavVol space available.

9 Rep Svc Source Filesystem Flow Resumed

Source SavVol space available.

10 Source Filesystem Frozen Result of specifying the autofreeze option or parameter.

11 Source Filesystem Thawed Release of autofreeze option or parameter.

12 Last Delta Replayed on Destination Filesystem

Issued at the end of a failover and reverse.

13 Redo Buffer close to overflow event In write-intensive environments, sometimes the redo buffers are not flushed fast enough.

14 Source mounted RO Result of specifying the autoro option or parameter.

15 Source mounted RW Release of autoro option or parameter.

Page 122: Celerra_Replcatr

Using Celerra Replicator (V1)122 of 172 Version 5.6.47

The events mentioned in Table 12 on page 122 apply to Celerra Replicator functionality.

VMCAST : Events for fs_copy (Facility ID : 84)

16 Replication in error Replication service is inactive or an internal error occurred.

Table 12 Events description

ID Description

0 Not an event

1 FS Copy over ip done

2 FS Copy over ip failed

3 Volume Copy over ip done

4 Volume Copy over ip failed

Table 11 Celerra Replicator events (page 2 of 2)

Event ID Event Description

Page 123: Celerra_Replcatr

123 of 172Version 5.6.47Using Celerra Replicator (V1)

Change the Celerra Replicator SavVol default size

By default, the system allocates 10 percent of the source file system size for each Celerra Replicator SavVol. You can enlarge this default amount by changing the value in the /nas/site/nas_param file.

Step Action

1. Log in to the Control Station.

2. Open /nas/site/nas_param with a text editor.

A short list of configuration lines appears.

!CAUTION!Do not edit anything in /nas/sys/, as these settings are overwritten with each code upgrade.

3. Locate the replication configuration line:

Replication:10:

where:10 = size (in %) of the source file system Celerra allocates for the replication SavVols.

The minimum SavVol size is 1 GB.

EMC recommends not using a value of less than 10%.

Note: Do not change any other lines in this file without a thorough knowledge of the potential effects on the system. Contact EMC Customer Service for guidance.

4. Change the parameter to represent the percentage of space you want to allocate.

5. Save and close the file.

Note: Changing this value does not require a Control Station restart.

Page 124: Celerra_Replcatr

Using Celerra Replicator (V1)124 of 172 Version 5.6.47

Change the passphrase between Celerra Network Servers

For a configuration with remote replication, use this procedure to change the passphrase on the Celerra system at each site.

Step Action

1. At each site, you can review the current passphrase using this command syntax:$ nas_cel -info <cel_name>

where:<cel_name> = name of Celerra Network Server

2. At each site, establish the new passphrase using this command syntax:$ nas_cel -modify <cel_name> -passphrase <passphrase>

where:<cel_name> = name of Celerra Network Server<passphrase> = new secure passphrase to be used for the connection, which must have 6- to 15-character and be the same on both sides of the connection

Example:# nas_cel -modify cs110 -passphrase nas_replication

operation in progress (not interruptible)...id = 5name = cs110owner = 503device =channel =net_path = 192.168.168.102celerra_id = APM000446038450000passphrase = nas_replication

Page 125: Celerra_Replcatr

125 of 172Version 5.6.47Using Celerra Replicator (V1)

Managing and avoiding IP replication problems

This section describes procedures to save replication sessions in the event of network, and source or destination file system outages, planned or unplanned. Additionally, this section details recommended practices to follow before performing replication. Failure to follow these recommendations might lead to nasdb inconsistencies and replication failures.

You can use Celerra Manager to perform much of the functionality described in this section. Consult Celerra Manager online help for instructions.

This section describes:

◆ "Preventive measures to avoid IP replication problems" on page 125

◆ "Replication restart methods" on page 128

◆ "Recovering from a corrupted file system" on page 130

◆ "Recovering from an inactive replication state" on page 135

◆ "Managing anticipated destination site or network outages" on page 131

◆ "Managing unanticipated destination site or network outages" on page 132

◆ "Managing unanticipated source site outages" on page 133

◆ "Managing expected source site outages" on page 133

◆ "Mount the destination file system read/write temporarily" on page 133

◆ "Recovering from an inactive replication state" on page 135

◆ "Creating checkpoints on the destination site" on page 136

◆ "Copy file system to multiple destinations with fs_copy" on page 136

Preventive measures to avoid IP replication problems

You can take the following preventive measures to avoid IP replication problems:

◆ "Creating restartable checkpoints for out-of-sync operations" on page 125

◆ "Controlling delta set size" on page 126

◆ "Enlarging SavVol size" on page 126

◆ "Calculating modifications rate on the source file system" on page 127

◆ "Accommodating network concerns" on page 127

Creating restartable checkpoints for out-of-sync operations

In any IP replication recovery scenario, it is critical to have restartable checkpoints as a base from which to work. Verifying that each file system contains checkpoints that are occasionally refreshed ensures that Celerra can perform an out-of-sync restart if required.

Restartable checkpoints created using fs_ckpt <srcfs> -name <srcfs>_repl_restart_1 -Create and fs_ckpt <srcfs> -name <srcfs>_repl_restart_2 -Create are automatically refreshed when starting from a differential copy. Restartable checkpoints are supported in version 5.4 only.

Page 126: Celerra_Replcatr

Using Celerra Replicator (V1)126 of 172 Version 5.6.47

Be sure to verify that these checkpoints are being refreshed by using fs_ckpt -list. For example, examine the creation times of the checkpoints, as shown in the following sample output:

$ fs_ckpt pfs3 -listid ckpt_name creation_time inuse full(mark) used258 pfs3_ckpt1 01/03/2006-04:20:10-EST y 90% 10%271 pfs3_ckpt2 01/03/2006-05:35:09-EST y 90% 10%282 pfs3_ckpt3 01/03/2006-06:22:07-EST y 90% 10%283 pfs3_ckpt4 01/03/2006-06:22:55-EST y 90% 10%284 pfs3_ckpt5 01/03/2006-06:34:26-EST y 90% 10%

If restartable checkpoints do not exist, create them. If they do exist, but their timestamps indicate they are not refreshing with the replication updates, check that the names are correct and replication is healthy.

"Out-of-sync replication relationship" on page 20 provides more information about these special checkpoints.

Controlling delta set size

Controlling the size of delta sets is integral to managing efficient replications. Version 5.4 and later enforce an 8 GB delta-set limit. Manageable delta sets are preferable to large deltas because:

◆ Large delta sets dramatically increase failover time (blockmap recovery).

◆ Large delta sets consume more operating system resources.

◆ Playback and create times for large delta sets do not rise proportionally with size. Reasonably sized delta sets are created and replayed faster.

Enlarging SavVol size

Before beginning replication, determine whether the default source side SavVol size (10 percent), as related to the file system size, is sufficient for the anticipated delta sets.

Find the SavVol size by calculating 10 percent of the file system size obtained by using nas_fs -size <fs_name>. If you previously changed the SavVol size, you can learn its present size using the fs_replicate -info <fs_name> command and checking the total_savevol_space field. To manage a large network outage or account for brief intervals when the incoming modification rate significantly exceeds the network’s ability to send changes to the destination site, you can increase the size of the replication SavVol. For example, a 500 GB file system that incurs 20 GB of change daily will, with a 50 GB SavVol, accommodate approximately two and one-half days of outage. If replication has already begun, change the SavVol size as follows:

Note: If you are starting replication, specify the SavVol size rather than use the default value.

Page 127: Celerra_Replcatr

127 of 172Version 5.6.47Using Celerra Replicator (V1)

Calculating modifications rate on the source file system

Estimate the rate of modifications expected in the source side file system. If the rate of anticipated change on the source file system is continuously greater than the available network bandwidth, the replication service cannot transfer data quickly enough to avoid becoming inactive. To avoid this state, create two restartable checkpoints (consult "Start replication without initiating a full data copy" on page 128). When autoro=yes the source file system becomes read only.

Accommodating network concerns

Consider the following quality of service (QoS) and network suggestions suitable to your bandwidth requirements:

◆ Use dedicated network devices for IP replication data transfer to avoid an impact on users.

◆ Apply QoS policy using a value that matches the bandwidth of one network to another. For instance, even if you have a 1 MB/s line from A to B and want to fill the pipe, set QoS at 1 MB/s so the line will not flood, causing packets to drop.

Configure the qos parameter to throttle bandwidth for a replication session (or bandwidths for different sessions) on a Data Mover using fs_replicate <fsname> -modify -option qos=<kbps>.

The EMC Celerra Network Server Command Reference Manual further details the fs_replicate command.

Step Action

1. Record existing policy parameters using this command syntax:$ fs_replicate -info <srcfs>

where:<srcfs> = name of the source file system

2. Stop replication using this command syntax:$ fs_replicate -suspend <srcfs> <dstfs>:cel=<cel_name>

where:<srcfs> = name of the source file system<dstfs> = name of the destination file system<cel_name> = name of the Celerra Network Server for the file system

3. Restart replication with the revised SavVol and old parameters using this command syntax:$ fs_replicate -restart <srcfs> <dstfs>:cel=<cel_name> savsize=<MB> -sav <srcsavvol_name> -option to=<value>, hwm=<MB>

where:<srcfs> = name of the source file system<dstfs> = name of the destination file system<cel_name> = name of the Celerra Network Server for the file systemsavsize=<MB> = size of the SavVol in MB<srcsavvol_name> = name of the source file system SavVol-option to=<value> = time-out interval in seconds-option hwm=<value> = high water mark in MB

Page 128: Celerra_Replcatr

Using Celerra Replicator (V1)128 of 172 Version 5.6.47

◆ If you do not expect significant packet loss, enable the fastRTO parameter. fastRTO determines which TCP timer to use when calculating retransmission timeout. The TCP slow timer (500 ms) is the default, causing the first time-out retransmission to occur in 1–1.5 seconds. But setting fastRTO to 1 sets the TCP fast timer (200 ms) for use when calculating retransmission timeout, which causes the first timeout to occur in 400–600 ms.

Use server_param <movername> -facility tcp fastRTO=1 to configure the setting.

Note: This setting may actually increase network traffic. Make the change cautiously, recognizing it might not improve performance.

◆ To ensure that a stable network transfer rate for delta-set transfers on a Data Mover, use a dedicated network port.

◆ Correctly set the TCP window size for network latency. Configuring the tcpwindow parameter sets the window size used by replication (and fs_copy). This value indicates the data load that can be sent before acknowledgment by the receiving site. Increasing the value is most effective with a high latency. Window size is calculated by multiplying the round-trip delay by the appropriate packet rate.

For example, to send 10 MB/s across an IP network with a round-trip delay of 100 ms, a window size of 1 MB (0.1 sec x 10 MB/s = 1 MB) is needed. Use server_param <movername> -facility rcp tcpwindow=bytes to configure the setting.

EMC Celerra Network Server Parameters Guide provides more information on setting tcp and rcp facilities.

Replication restart methods

If file systems are corrupted or out-of-sync, you can:

◆ "Start replication without initiating a full data copy" on page 128

◆ "Start replication from scratch" on page 129

You must start replication from scratch if the source or destination file system is corrupted. If file systems are only out-of-sync, you can restart the replication relationship.

Note: If you restart a replication session and there are unmounted checkpoints, a full data copy will be initiated instead of a differential copy.

Start replication without initiating a full data copy

"Restarting a replication relationship" on page 89 provides details on how to start a replication session without having to perform a full data copy.

Page 129: Celerra_Replcatr

129 of 172Version 5.6.47Using Celerra Replicator (V1)

Start replication from scratch

Step Action

1. Terminate the replication relationship using this command syntax:$ fs_replicate -abort <srcfs> <dstfs>:cel=<cel_name>

where:<srcfs> = name of the source file system<dstfs> = name of the destination file system

<cel_name> = name of Celerra Network Server for the file system

2. Create a checkpoint of the source file system using this command syntax:$ fs_ckpt <srcfs> -Create

where:<srcfs> = name of the source file system

3. Copy the checkpoint to the destination file system using this command syntax:$ fs_copy -start <src_ckpt> <dstfs>:cel=<cel_name>

where:<src_ckpt> = source checkpoint copied of the destination<dstfs> = name of the destination file system<cel_name> = name of the destination Celerra Network Server

4. Convert the destination file system to rawfs using this command syntax:$ nas_fs -Type rawfs <dstfs> -Force

where:<dstfs> = name of the destination file system

5. Start replication from the source to destination file system using this command syntax:$ fs_replicate -start <srcfs> <dstfs>:cel=<cel_name>

where:<srcfs> = name of the source file system<dstfs> = name of the destination file system<cel_name> = name of the destination Celerra Network Server

6. Create a second checkpoint of the source file system using this command syntax:$ fs_ckpt <srcfs> -Create

where:<srcfs> = name of the source file system

7. Perform a differential copy by typing:$ fs_copy -start <src_newckpt> <dstfs>:cel=<cel_name> -fromfs <previous_ckpt> monitor=off

where:<dstfs> = name of the destination file system<cel_name> = name of the destination Celerra Network Server

8. Check the copy’s progress and completion using this command syntax:$ fs_copy -info <srcfs> or fs_copy -list

where:<srcfs> = name of the source file system

Page 130: Celerra_Replcatr

Using Celerra Replicator (V1)130 of 172 Version 5.6.47

Recovering from a corrupted file system

To recover from a corrupted file system:

◆ "Run nas_fsck on the source file system" on page 130

◆ "Suspend and restart replication" on page 130

Note: You cannot use nas_fsck if the destination file system is corrupted as a result of an improper fs_copy operation. File system replication fails due to pending nas_fsck.

Run nas_fsck on the source file system

The nas_fsck command checks and repairs a file system when replication is threatened. You can run the command to check the source file system while replication is running. If nas_fsck detects inconsistencies in the primary file system, the changes that occur as a result of using nas_fsck are replicated to the destination file system like any other file system modifications.

Suspend and restart replication

If the destination is so unstable or corrupted that it might fail before the nas_fsck changes are replicated and replayed, suspend replication, run nas_fsck -start <srcfs>, and restart replication. While it is unlikely that only the destination file system would be corrupted, you can diagnose a destination-only failure on the

Step Action

1. Modify time-out and HWM values to zero using this command syntax:$ fs_replicate -modify <srcfs> -option hwm=0,to=0

where:<srcfs> = name of the source file system

Note: Setting trigger points to zero causes Celerra to keep replication active and track changes, but not cut delta sets.

2. Run nas_fsck on the source file system to replicate and replay changes on the destination file system using this command syntax:$ nas_fsck -start <srcfs>

where:<srcfs> = name of the source file system

Note: Running nas_fsck repairs corruption on the source file system, bringing it into a consistent, but not original, state. While nas_fsck runs, the file system is not mounted to avoid system instability. When the command is complete and inconsistencies addressed, the file system is brought back online.

3. Revert to your previous time-out and HWM values using this command syntax:$ fs_replicate -modify <srcfs> -option hwm=<MB>, to=<second>

where:<srcfs> = name of the source file systemhwm=<value> = original high water mark value in MBto=<value> = original time-out interval in seconds

Page 131: Celerra_Replcatr

131 of 172Version 5.6.47Using Celerra Replicator (V1)

source file system using the nas_fsck -start <srcfs> command. Any inconsistencies found by nas_fsck on the primary file system are replicated to the secondary file system.

Managing anticipated destination site or network outages

This section describes how to handle expected destination or network outages. When planning for an outage, for instance, to restart a secondary Data Mover or to conduct standard network maintenance involving the secondary file system, follow the guidelines in the following sections to protect replication:

◆ "Anticipated destination site outage over a short period" on page 131

◆ "Anticipated destination site outage over a long period" on page 131

Anticipated destination site outage over a short period

Begin by evaluating the outage period and whether the site can survive it. If the data queues easily in the SavVol, nothing need be done. For example, if the planned outage period is one day, the SavVol is 100 MB, the file system is 1 GB, and 200 MB of modifications occur daily, then survival is ensured for a half-day because the SavVol will fill in 12 hours. On the other hand, if only 100 MB in modifications occur daily, a whole day’s worth of changes are protected.

Anticipated destination site outage over a long period

In situations in which an outage is expected for a long period and the replication service continues running, the SavVol might become full or trigger flow control, eventually leading replication to drop out-of-sync.

If this scenario is likely, perform a replication suspend and restart, as directed in the following procedure.

Step Action

1. Suspend replication using this command syntax:$ fs_replicate -suspend <srcfs> <dstfs>:cel=<cel_name>

where:<srcfs> = name of the source file system<dstfs> = name of the destination file system<cel_name> = name of Celerra Network Server for the file system

2. Run nas_fsck using this command syntax at the source site:$ nas_fsck -start <srcfs>

where:<srcfs> = name of the source file system

3. Restart the replication using this command syntax:$ fs_replicate -restart <srcfs> <dstfs>:cel=<cel_name>

where:<srcfs> = name of the source file system<dstfs> = name of the destination file system<cel_name> = name of Celerra Network Server for the file system

Page 132: Celerra_Replcatr

Using Celerra Replicator (V1)132 of 172 Version 5.6.47

Note: To restart after suspending replication, you must use the -restart option.

Managing unanticipated destination site or network outages

This section describes coping with unanticipated Celerra destination platform or network outages. Most destination outages require no action. For instance, if Celerra restarts or goes offline—and the system does not fall out-of-sync—no remediation is necessary. Consider the following:

◆ If fs_replicate -list shows replication is still active after the unplanned destination outage is finished, nothing need be done.

◆ If fs_replicate -list shows replication is inactive and out-of-sync after the unplanned destination outage is finished, consult "Replication restart methods" on page 128.

Step Action

1. Suspend each replication session individually using this command syntax:$ fs_replicate -suspend <srcfs> <dstfs>:cel=<cel_name>

where:<srcfs> = name of the source file system<dstfs> = name of the destination file system<cel_name> = name of the destination Celerra Network Server

Note: The suspend operation lets Celerra track changes in a SavVol that automatically expands. Less total storage is needed because tracking is done only once in the checkpoint SavVol, not twice in the checkpoint and replication SavVol. Also, checkpoints retain only one overall changed block instead of one per delta set.

2. When all replication sessions are suspended, to check that the session you suspended no longer appears in output for fs_replicate -list, type:$ fs_replicate -list

Note: Use this command on source and destination Celerra systems to verify that no sessions are running.

Output from the command should display no sessions, as shown below: Local Source FilesystemsId Source FlowCtrl State Destination FlowCtrl State Network

3. Verify the size of the suspend checkpoint to ensure that there is enough disk space to expand the SavVol. The suspend checkpoint, root_suspend_ckpt, is used to restart replication and is added by the -suspend option. Verify the size by typing:$ nas_fs -size root_suspend_ckpt

4. Restart replication after the outage is over using this command syntax:$ fs_replicate -restart <srcfs> <dstfs>:cel=<cel_name>

where:<srcfs> = name of the source file system<dstfs> = name of the destination file system<cel_name> = name of the destination Celerra Network Server

Page 133: Celerra_Replcatr

133 of 172Version 5.6.47Using Celerra Replicator (V1)

◆ During an unplanned destination outage, the only way to mitigate the system impact other than restoring the destination file system before the SavVol fills up is to reconfigure a small timeout and HWM on Celerra, if applicable. In this case, use fs_replicate -modify -option to=0, hwm=8192 to save a small amount of SavVol space, allowing the system more time before falling out-of-sync. This causes delta sets to be cut as large and with as little duplication as possible. Contrast a TO of zero with a TO of 60 seconds, which consumes 128 MB of SavVol per minute even without many modifications because the minimum delta-set size is 128 MB.

Note: If replication failed and corrupted the destination file system, or you mistakenly suspended replication, mounted the destination read/write, and restarted replication, you must abort the session and restart using a full fs_copy. See "Replication restart methods" on page 128 for more information.

Managing unanticipated source site outages

This section describes how to manage unplanned Celerra source outages such as power interruptions or restarts.

Consider the following if you decide to activate the Data Recovery (DR) site:

◆ "Replication restart methods" on page 128 describes the source that resides on a CLARiiON system that lost power, and replication is inactive.

◆ If the source resides on a Symmetrix system that lost power—where replication should still be active—no remediation is necessary.

◆ If only the source is down and you do not want to activate the DR site, no remediation is necessary.

◆ "Recovering replication data" on page 61 describes replication failover, resynchronization, and reversal.

Managing expected source site outages

This section describes coping with both anticipated and unexpected outages on the source Celerra. Perform the same steps as described in "Managing unanticipated source site outages" on page 133, and decide whether to activate the DR site.

Mount the destination file system read/write temporarily

This section describes temporarily mounting the secondary file system as a read/write volume to test data recovery without doing a failover and taking the source server offline.

To temporarily mount the destination read/write, perform this procedure.

Page 134: Celerra_Replcatr

Using Celerra Replicator (V1)134 of 172 Version 5.6.47

!CAUTION!The following procedure cannot be used for VDM root file systems. A checkpoint of a VDM’s root file system is designed in such a way that it cannot be restored. Replication for VDM root file system needs to be restarted with a full fs_copy to ensure that the source and destination file systems are synchronized. Failure to do this will lead to file system corruption and data unavailability.

Step Action

1. Suspend replication using this command syntax:$ fs_replicate -suspend <srcfs> <dstfs>:cel=<cel_name>

where:<srcfs> = name of the source file system<dstfs> = name of the destination file system<cel_name> = name of Celerra Network Server for the file system

2. Assign the file system type to default using this command syntax:$ nas_fs -Type uxfs <dstfs> -Force

where:<dstfs> = name of the destination file system

3. Unmount the destination file system using this command syntax:$ server_umount <movername> -perm <dstfs>

where:<movername> = name of the Data Mover<dstfs> = name of the destination file system

4. Mount the destination file system read-write using this command syntax:$ server_mount <movername> -option rw <dstfs> /<dstfs_mountpoint>

where:<movername> = name of the Data Mover<dstfs> = name of the destination file system

<dstfs_mountpoint> = point at which the file system is mounted

5. Perform DR testing with an appropriate program after the both sites are read-write. A number of diagnostics help to ensure that your database can start up correctly, such as:

• Reading/writing to a single file or every file• Creating new files• Modifying existing files• Reading or deleting file systems

6. Make the now-writable destination file system available to clients either by exporting it through NFS, or sharing it through CIFS (assuming a CIFS server is active on the destination side) using this command syntax:$ server_export <movername> -Protocol nfs -option <options> <pathtoexport> or$ server_export <movername> -Protocol cifs -name <sharename> -option <options> <pathtoshare>

where:<movername> = name of the Data Mover

Page 135: Celerra_Replcatr

135 of 172Version 5.6.47Using Celerra Replicator (V1)

Recovering from an inactive replication state

If the state of replication and playback is inactive due to a power failure on Celerra or because flow control was exercised—and source and destination file systems are out-of-sync—you must restart the replication relationship. Go to "Replication restart methods" on page 128.

7. Unmount the destination file system after testing ends using this command syntax:$ server_umount <movername> -perm <dstfs>

where:<movername> = name of the Data Mover<dstfs> = name of the destination file system

8. Mount the destination file system as read-only using this command syntax:$ server_mount <movername> -option ro <dstfs> /<dstfs_mountpoint>

where:<movername> = name of the Data Mover<dstfs> = name of the destination file system<dstfs_mountpoint> = point at which the file system is mounted

9. Convert the destination file system to rawfs after the restore ends using this command syntax:$ nas_fs -Type rawfs <dstfs> -Force

where:<dstfs> = name of the destination file system

10. Restart replication at the source site using this command syntax:$ fs_replicate -restart <srcfs> <dstfs>:cel=<cel_name>

where:<srcfs> = name of the source file system <dstfs> = name of the destination file system<cel_name> = name of Celerra Network Server for the file system

Step Action

Page 136: Celerra_Replcatr

Using Celerra Replicator (V1)136 of 172 Version 5.6.47

Creating checkpoints on the destination site

When using the fs_copy command to regularly copy modifications from source to destination file system, you should employ checkpoints on the destination site for these reasons:

◆ While an fs_copy is in progress, the file system on the destination is in a rawfs state and unavailable. However if you preserve a checkpoint on the destination, this checkpoint remains available during the copy. "Copy file system to multiple destinations with fs_copy" on page 136 provides more details.

◆ By using checkpoints, you can preserve the copied views of the file system on the destination site. When checkpointing the destination, if an fs_copy is in progress, the operation suspends until the copy is complete.

The following caveats apply when creating checkpoints on the destination file system with replication running:

◆ Checkpoints created on the destination file system are supported just as on the source file system, except that they fall on delta-set boundaries.

◆ Checkpoints processed during playback suspend until the playback ends. So, if you are in the midst of a large delta-set playback, the refresh or create could take an extended period to process while Celerra waits for the playback to finish.

Copy file system to multiple destinations with fs_copy

The fs_copy command is used typically in a first-time replication start to manually synchronize source and destination file systems. It is also often employed in a script to copy one source file system to many destinations, either by cascading one file system to a number of destinations or directly to each location.

Using fs_copy does not constitute replication. It is a copy command that emulates replication which requires considerable scripting and integration. The command should be used sparingly by a knowledgeable operator.

The following scenario describes regularly copying a file system to three destination sites (A, B, and C). The procedure requires an operator to perform differential copies on alternating checkpoint sets.

Note: It is highly recommended that slice volumes be used to create and extend file systems for this fs_copy solution. Managing EMC Celerra Volumes and File Systems with Automatic Volume Management provides more information about slice volumes.

Page 137: Celerra_Replcatr

137 of 172Version 5.6.47Using Celerra Replicator (V1)

Step Action

1. Copy a file system to multiple destinations with fs_copy and create a checkpoint of the source file system using this command syntax:$ fs_ckpt <srcfs> -name <src_ckpt1> -Create

where:<srcfs> = name of the source file system<src_ckpt1> = first checkpoint on source

Note: If the primary file system extends before checkpoint creation, the fs_copy command will fail. In this case, you must extend the destination file system manually to keep file system sizes identical. First convert the file system to rawfs using the nas_fs -Type rawfs command. Then use the nas_fs -xtend command. If a slice volume is not used, an incremental copy might fail and you might need to run a full fs_copy from scratch when the source file system is extended.

2. Copy it to the destination A file system after the checkpoint is created using this command syntax:$ fs_copy -start <src_ckpt1> <dstfs>:cel=<cel_name>

where:<src_ckpt1> = first checkpoint on source<dstfs> = destination name for file system A<cel_name> = name of Celerra Network Server for file system A

Note: While an fs_copy is running, the destination file system is inaccessible. To make it accessible, you must create a checkpoint before the data transfer starts.

3. After the fs_copy operation ends, create a checkpoint of destination file system using this command syntax:$ fs_ckpt <dstfs> -name <dst_ckpt1> -Create

where:<dstfs> = destination file system name A<dst_ckpt1> = first checkpoint on destination file system A

4. Create a second checkpoint of the source file system by typing:$ fs_ckpt <srcfs> -name <src_ckpt2> -Create

where:<srcfs> = name of the source file system<src_ckpt2> = second checkpoint on source

5. Convert destination A file system to rawfs using this standard syntax:$ nas_fs -Type rawfs <dstfs> -Force

where:<dstfs> = destination name of file system A

Page 138: Celerra_Replcatr

Using Celerra Replicator (V1)138 of 172 Version 5.6.47

6. Perform a differential copy between checkpoints 1 and 2 using this command syntax:$ fs_copy -start <src_ckpt2> <dstfs>:cel=<cel_name> -fromfs <src_ckpt1>

where:<src_ckpt2> = second checkpoint on the source<dstfs> = destination name for file system A<cel_name> = name of Celerra Network Server for file system A<src_ckpt1> = first checkpoint on the source

7. After the fs_copy command completes again, refresh the first checkpoint of destination file system A using this command syntax:$ fs_ckpt <dst_ckpt1> -refresh

where:<dst_ckpt1> = first checkpoint on destination file system A

8. Perform steps 2 through 7 for destinations B and C. The source file system is now saved on destinations A, B, and C.

9. Refresh the copy of the source file system at destinations A, B, and C, refresh the first source checkpoint using this command syntax:$ fs_ckpt <src_ckpt1> -refresh

where:<src_ckpt1> = first checkpoint on source

10. Repeat steps 2 through 8, swapping checkpoints <src_ckpt1> and <src_ckpt2>.

Step Action

Page 139: Celerra_Replcatr

139 of 172Version 5.6.47Using Celerra Replicator (V1)

Transporting replication data using disk or tape

Copying the baseline source file system from the source to the destination site over the IP network can be a time-consuming process. You can use an alternative method by copying the initial checkpoint of the source file system, backing it up to a disk array or tape drive, and transporting it to the destination site, as shown in Figure 7 on page 139.

Figure 7 Physical transport of data

Note: Use the IP network to make the initial copy of the root file system for a VDM.

This section consists of the following tasks:

◆ "Disk transport method" on page 140

◆ "Tape transport method" on page 144

Destination site

Celerra DataMover

Source site

CLARiiON orSymmetrix

storage system

CLARiiON orSymmetrix

storage system

Celerra DataMover

Truck

Filesystem

Filesystem

Disk ortape

Disk ortape

CNS-000767

Page 140: Celerra_Replcatr

Using Celerra Replicator (V1)140 of 172 Version 5.6.47

Disk transport method

If the source file system holds a large amount of data, the initial copy of the source to the destination file system can be time-consuming to move over the IP network. Preferably, move the initial file system copy by disk, instead of over the network.

If you want to physically transport a copy of your source file system using a CLARiiON disk array, the Data Movers that run replication must connect to the storage system using Fibre Channel switched fabric connections.

To transport the baseline copy of the source file system to the destination site, use the disk transport method described in this section or contact EMC about a customized replication baselining service offering.

Note: You can use any qualified CLARiiON storage system for this transfer.

To transport replication data using disk:

1. "Capture data from the source site on disk" on page 141

2. "Transfer data to the destination site from disk" on page 143

Page 141: Celerra_Replcatr

141 of 172Version 5.6.47Using Celerra Replicator (V1)

Step 1: Capture data from the source site on disk

Step Action

1. List the disks attached to the Celerra Network Server using the nas_disk -list command. Keep this list to use for comparison later in this procedure.

2. Attach a supported CLARiiON array (for example a CX300) with the appropriately bound LUNs to the Celerra Network Server.

This procedure assumes you will use a dedicated array. "Setting up the CLARiiON disk array" on page 147 describes preparing the CLARiiON array.

3. Probe and verify SCSI disks by typing:$ server_devconfig server_2 -probe -scsi -disks

where server_2 is the Data Mover with access to the CLARiiON array.

4. Create SCSI disks by typing:$ nas_diskmark -mark -all

The disks will be available to all Data Movers including standby Data Mover.

5. List the disks that are attached to the Celerra Network Server using the nas_disk -list command. Then perform a diff command between this list and the one created in step 1 of "Capture data from the source site on disk" on page 141. For example:

> 378 n 260607 APM00034402893-0000 CLSTD d378 2> 379 n 260607 APM00034402893-0001 CLSTD d379 2

6. Create a file system on the CLARiiON array that is the same size as the source file system. To do so:

a. Create a user-defined pool by typing:$ nas_pool -create -name transport_disks -volumes d378,d379

b. Create a file system by typing: $ nas_fs -name transport_fs -type rawfs -create samesize=src pool=transport_disks -option mover=server_2

c. Create a mountpoint for the file system.d. Mount the file system read-only.

Note: Ensure that you create the file system as rawfs and use the samesize= option to ensure that it is identical in size to the source file system.When creating the pool, ensure that the disks are added in the same order on source and destination sites. If you are creating more than one file system, ensure that they are created in the same order on source and destination sites. EMC recommends that you create the largest file system first.

7. Create a checkpoint of the source file system by typing:$ fs_ckpt src -Create

8. Copy the source file system checkpoint to the file system created on the new disks by typing:$ fs_copy -start src_ckpt1 transport_fs -option convert=no, monitor=off

Page 142: Celerra_Replcatr

Using Celerra Replicator (V1)142 of 172 Version 5.6.47

9. Monitor the fs_copy progress by typing:

$ fs_copy -list

$ fs_copy -info <session_id>

Verify the size of the transport file system to the source file system using the nas_fs -size command.

10. Delete the disks from the source site. To do so:

a. Unmount the file system transport_fs (server_umount)

b. Delete the mountpoint (server_mountpoint)c. Delete the file system (nas_fs -delete)d. Delete the pool (nas_pool -delete)e. Delete the disks which were the result of the -diff command in step 5 (nas_disk -delete)

of "Capture data from the source site on disk" on page 141

11. Verify that the disks were removed by using the nas_disk -list command. The results you obtain from this step should be the same as those derived from the first step of "Capture data from the source site on disk" on page 141.

12. Disconnect and uninstall the CLARiiON array from the source site.

13. Transport the disk array to the destination site.

Step Action

Page 143: Celerra_Replcatr

143 of 172Version 5.6.47Using Celerra Replicator (V1)

Step 2: Transfer data to the destination site from disk

Step Action

1. List the disks that are attached to the Celerra Network Server using the nas_disk -list command. Keep this list to use for comparison later in this procedure.

2. Attach the CLARiiON array at the destination site.

3. Probe and verify SCSI disks by typing:$ server_devconfig server_2 -probe -scsi -disks

where server_2 is the Data Mover with access to the CLARiiON array.

4. Create SCSI disks by typing:$ nas_diskmark -mark -all

The disks will be available to all Data Movers including standby Data Mover.

5. List the disks that are attached to the Celerra Network Server using the nas_disk -list command. Then perform a diff command between this list and the one created in step 1 of "Transfer data to the destination site from disk" on page 143. For example:

> 375 n 260607 APM00034402893-0000 CLSTD d375 1> 376 n 260607 APM00034402893-0001 CLSTD d376 1

6. Create a file system on the transport disk array. To do so:

a. Create a user-defined pool by typing:$ nas_pool -create -name transport_disks -volumes d375,d376

b. Create a file system by typing: $ nas_fs -name transport_fs -type rawfs -create samesize=src:cel=eng16853 pool=transport_disks -option mover=server_2

c. Create a mountpoint.d. Mount the file system read-only.

Note: Ensure that you create the file system as rawfs and use the samesize= option to ensure that it is identical in size to the source file system. When creating the pool ensure that the disks are added in the same order on source and destination sites. If you are creating more than one file system ensure that they are created in the same order on source and destination sites.

7. Create a destination file system (in this example the destination Celerra file system is attached to a Symmetrix storage system). To do so:

a. Create a file system by typing: $ nas_fs -name dest -type rawfs -create samesize=src:cel=eng16853 pool=symm_std

b. Create a mountpoint and mount the file system.

Note: Ensure that you create the file system as rawfs and use the samesize= option to ensure that it is identical in size to the source file system.

8. Copy the file system on the transport disk array to the destination file system, created in step 7 of "Transfer data to the destination site from disk" on page 143:$ fs_copy -start transport_fs dest -option convert=no

Page 144: Celerra_Replcatr

Using Celerra Replicator (V1)144 of 172 Version 5.6.47

Tape transport method

If the source file system contains a large amount of data, the initial copy of the source file system to the destination file system can be time-consuming to move over the IP network. Moving the initial copy of the file system by backing it up to tape, instead of over the network, is preferable.

When using this method of transport, note the following:

◆ You must have a valid NDMP infrastructure on both Celerra Network Servers.

◆ The restore is performed to a rawfs file system that must be mounted on the Data Mover.

◆ The restore will be rejected if the destination file system is not the same size as the source file system used for the backup.

Note: This special backup is used only for transporting replication data.

!CAUTION!Backing up file systems from a Unicode-enabled Data Mover and restoring to an ASCII-enabled Data Mover is not supported. I18N mode (Unicode or ASCII) must be the same on the source and destination Data Movers.

To transport replication data using tape:

1. "Capture data from the source site on tape" on page 145

2. "Transfer data to the destination site from tape" on page 146

9. The destination file system is now rawfs and contains a copy of the source file system checkpoint.

10. Delete the disks from the transport disk array. To do so:

a. Unmount the file system transport_fs (server_umount)

b. Delete the mountpoint (server_mountpoint)c. Delete the file system (nas_fs -delete)d. Delete the pool (nas_pool -delete)e. Delete the disks which were the result of the -diff command in step 5 (nas_disk -delete)

of "Transfer data to the destination site from disk" on page 143.

11. Verify that the disks were removed by using the nas_disk -list command. The results you obtain from this step should be the same as those derived from step 1 of "Transfer data to the destination site from disk" on page 143.

12. Disconnect and uninstall the CLARiiON array from the destination site.

13. Continue with the next step for setting up replication, Task 6: "Begin replication" on page 48, in Using EMC Celerra Replicator (V1).

Step Action

Page 145: Celerra_Replcatr

145 of 172Version 5.6.47Using Celerra Replicator (V1)

Step 1: Capture data from the source site on tape

Step Action

1. Create the checkpoint of the source file system by typing:$ fs_ckpt src_ufs1 -Create

Note: You can back up only the checkpoint of an IP replication read-only target file system using the NDMP backup feature. If you attempt to back up a replication read-only target file system, NDMP will fail when replication is updating the change. Celerra Network Server version 5.5.27 and later supports NDMP backup of integrated checkpoints and manually created checkpoints of a target replication file system.

2. Set the NDMP environment variable for your backup software. For example, set the VLC=y NDMP environmental variable before you run the backup.

The NDMP technical module for your particular backup software provides information about environment variables. For information about how to set this variable, read your backup software vendor’s documentation.

Note: The source file system and the checkpoint must be mounted on the NDMP Data Mover.

3. Use your normal backup procedure to back up the source file system checkpoint.

4. Transport the backup tapes to the destination site.

Page 146: Celerra_Replcatr

Using Celerra Replicator (V1)146 of 172 Version 5.6.47

Step 2: Transfer data to the destination site from tape

Step Action

1. When the tapes are on the destination Celerra Network Server, create a file system (rawfs) that is the same size as the source file system. Create the file system on a metavolume, create a mount point, and then mount the file system. Managing EMC Celerra Volumes and File Systems Manually describes how to create a file system.

Ensure that you create the file system as rawfs and use the samesize= option to ensure that it is identical in size to the source file system.

2. Determine the volume number of the destination file system created in step 1. In this example, the volume number for the rawfs file system is 66:$ nas_fs -list

id inuse type acl volume name server1 y 1 0 66 rawfs server_22 y 1 0 68 new server_33 y 1 0 70 last server_4

3. Using your normal NDMP restore procedure, restore the backup using the following as the file system name:/.celerra_vol_<fs_volume_ID>

where:<fs_volume_ID> = volume number of the rawfs file system (66 in the example)

Note: The file system must be restored to the NDMP Data Mover.

4. The destination file system is now rawfs and contains the source file system checkpoint.

Start the replication between source and destination. Follow the procedure, Task 6: "Begin replication" on page 48, to set up remote replication.

5. Create a second checkpoint. Follow the procedure, Task 7: "Create a second checkpoint of the source file system" on page 50.

6. Perform an incremental copy and allow the destination system to convert to uxfs. Follow the procedure, Task 8: "Copy incremental changes" on page 52.

Be sure to specify the -force option, for example:

$ fs_copy -start src_ufs1_ckpt2 dst_ufs1:cel=cs110 -fromfs src_ufs1_ckpt1 -Force -option monitor=off

7. Verify the replication session:

$ fs_replicate -list

Page 147: Celerra_Replcatr

147 of 172Version 5.6.47Using Celerra Replicator (V1)

Setting up the CLARiiON disk array

If you want to physically transport a copy of your source file system using a CLARiiON disk array, you must set up replication on a Celerra gateway.

Use the following documents in this setup procedure:

◆ EMC CLARiiON CX300, CX500, and CX700 Storage Systems Initialization Guide

◆ EMC CLARiiON CX300 2-Gigabit Disk Processor Enclosure (DPE2) Setup and Cabling Guide

Note: Use the appropriate setup and cabling guide depending on the disk array used.

When configuring the CLARiiON disk array for transporting replication data, run the appropriate setup script after setting up zoning for the network switches.

To prepare the CLARiiON disk array to receive the copy of the source file system:

◆ "Review the prerequisites" on page 147

◆ "Run the setup script" on page 149

◆ "Create data LUNs" on page 151

Review the prerequisites

Step Action

1. Cable and zone the CXxxx disk array to the Celerra Network Server.

2. Ensure that the required software components are installed on the CLARiiON disk array:

• CXxxx Base Array (EMC FLARE®)• EMC Navisphere®_ArrayAgent • Navisphere Management UI• EMC Access Logix™

Read the E-Lab Interoperability Navigator for the most recent Celerra software and FLARE microcode compatibility specifications.

3. The E-Lab Interoperability Navigator, which is available at Powerlink as definitive information on supported software and hardware, such as backup software, Fibre Channel switches, and application support for Celerra network-attached storage (Celerra) products.

The E-Lab Interoperability Navigator is for EMC use only. Do not share this information with customers.

Page 148: Celerra_Replcatr

Using Celerra Replicator (V1)148 of 172 Version 5.6.47

4. Make sure you have the following:

• Network information (hostname, IP address, subnet mask, and gateway for the SPs).• Set up the service computer with:

• Windows Server 2000 or Windows NT 4.0 with Service Pack 5 (or later).• Dial-up networking using direct connection.

• The latest version of the Navisphere CLI is installed in C:\program files\emc\navisphere cli on the service computer.

• Null modem cable with 9-pin female-to-female connectors.• An IP connection between the Celerra Network Server and the SPs (required for

installing the system software).

5. Create the PPP link by adding a modem and creating a connection as described in the EMC CLARiiON CX300, CX500, and CX700 Storage Systems Initialization Guide.

Step Action

Page 149: Celerra_Replcatr

149 of 172Version 5.6.47Using Celerra Replicator (V1)

Run the setup script

Step Action

1. Using the NULL modem cable, connect the service computer to the SP A serial port.

2. Establish a dial-up connection between the service computer and SP A by selecting, Start > Settings > Network and Dial-up Connections > Direct Connection.

The Connect Direct Connection dialog box opens.

3. Establish a dial-up connection by filling in the dialog box as follows:

Username: clariion

Password: clariion!

4. Select Connect.

Note: Establishing this connection might require several redial attempts.

5. Open a Command Prompt Window by selecting Start > Programs > Accessories > Command Prompt.

The Command Prompt window opens.

6. Insert the Celerra Installation CD into the CD-ROM drive.

7. In the Command Prompt window, change to the \clariion directory on the CD-ROM by typing the drive letter and then typing cd \clariion.

For example, type:

D:cd\clariion

Serial port

CNS-000760

Page 150: Celerra_Replcatr

Using Celerra Replicator (V1)150 of 172 Version 5.6.47

8. Run the script by typing setup, and press Enter:=========================================================== Program commandline: setup Startup values: Version: 7.0 Jumppoint into program: Pathname to navicli: “c:\program files\emc\navisphere cli\navicli” Debug (1=yes) : Logging (1=yes): 1 Logfile: C:\DOCUME~1\ADMINI~1\LOCALS~1\Temp\laptop_script.log IP Address for PPP: 128.221.252.1 Commands will all be sent via the dial-up connection.

Disconnect any existing Dial-Up Networking connections. Connect serial cable to SPA and start Dial-up Networking. When the connection is established, press any key to continue configuring the array. Connection to SPA (128.221.252.1) has been established and verified.

Waiting to ensure SP is up and stable ... NOTE: Some operations, like changing the serial number, result in multiple reboots. This is expected.

Checking 128.221.252.1 SP at 128.221.252.1 has responded. Waiting 60 seconds to ensure it stays up. SP at 128.221.252.1 is up.

Configuring CX Series ...

Creating the hotspare ...

Setting Failover Mode

When prompted to create Celerra Control Volumes, type N:

Do you want to configure Celerra Control Volumes? [Y,N,] : N

Configuring cache

Configuring the cache ...

Cache Configured.

Tasks complete. The serial connection to the SP can now be disconnected.

This procedure was logged into file: C:\DOCUME~1\ADMINI~1\LOCALS~1\Temp\laptop_script.log

When the script completes, type exit to close the Command Prompt window.

Step Action

Page 151: Celerra_Replcatr

151 of 172Version 5.6.47Using Celerra Replicator (V1)

Create data LUNs

Step Action

1. Log in to the Celerra Network Server.

Change to root (su) and type the password.

2. Determine the CXxxx serial number from a DOS window by typing:$ navicli -h <SPA_IP_address> getagent

Output:

Agent Rev: 6.6.0 (3.1)Name: K10Desc:Node: A-APM00035106458Physical Node: K10Signature: 979596Peer Signature: 885857Revision: 2.06.500.4.004SCSI Id: 0Model: 500Model Type: RackmountProm Rev: 3.00.00SP Memory: 2048Serial No: APM00035106458SP Identifier: ACabinet: DPE2

Page 152: Celerra_Replcatr

Using Celerra Replicator (V1)152 of 172 Version 5.6.47

3. Run the setup script using this command syntax:

$ /nas/sbin/setup_clariion <CXxxx_serial_number>Use the serial number from the previous step.

Output:

# nas//setup_clariion APM00035106458

CLARIION(s) APM00035106458 will be setup.

Setup CLARiiON APM00035106458 storage device...

Enter the ip address for A_APM00035106458: 172.24.168.72

Enter the ip address for B_APM00035106458: 172.24.168.73System 172.24.168.72 is upSystem 172.24.168.73 is up

Clariion Array: APM00035106458 Model: CX500 Memory: 2048

Committed Base software package

The following 5 template(s) available:

1. CX_All_4Plus1_Raid_52. CX_Standard_Raid_53. CX_Standard_Raid_14. CX_Standard_Raid_5_Legacy5. CX_Standard_Raid_1_LegacyPlease select a template in the range of 1-5 or 'q' to quit: 1Summary: 2 disk group(s) are created. 8,9 5 spare(s) are created. 200,201,202,203,204Enclosure(s) 0_0 are installed in the system.Enclosure info:---------------------------------------------------------------- 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14----------------------------------------------------------------0_0: 146 146 146 146 146 146 146 146 146 146 146 146 146 146 146 *8 *8 *8 *8 *8 *HS *9 *9 *9 *9 *9 *HS *HS *HS *HS----------------------------------------------------------------"*" indicates a diskgroup/spare which will be configuredSize Type Disks Spares------------------------- 146 FC 15 5

Do you want to continue and configure as shown [yes or no]?: yes

Enclosure 0_0.Created disk group 8, luns 16,17Created spare 200Created disk group 9, luns 18,19Created spare 201Created spare 202Created spare 203Created spare 204

Binding complete.All luns are created successfully!

Enclosure(s) 0_0 are installed in the system.

Step Action

Page 153: Celerra_Replcatr

153 of 172Version 5.6.47Using Celerra Replicator (V1)

3. Enclosure info:---------------------------------------------------------------- 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14----------------------------------------------------------------0_0: 146 146 146 146 146 146 146 146 146 146 146 146 146 146 146 8 8 8 8 8 HS 9 9 9 9 9 HS HS HS HS----------------------------------------------------------------

Configuration completed!

Setup of CLARiiON APM00035106458 storage device complete.

4. Register the World Wide Names using the Connectivity Status window in Navisphere.

5. Proceed with "Disk transport method" on page 140 procedure.

Step Action

Page 154: Celerra_Replcatr

Using Celerra Replicator (V1)154 of 172 Version 5.6.47

Troubleshooting Celerra Replicator

As part of an effort to continuously improve and enhance the performance and capabilities of its product lines, EMC periodically releases new versions of its hardware and software. Therefore, some functions described in this document might not be supported by all versions of the software or hardware currently in use. For the most up-to-date information on product features, refer to your product release notes.

If a product does not function properly or does not function as described in this document, please contact your EMC Sales Representative.

Where to get help

Product information – For documentation, release notes, software updates, or for information about EMC products, licensing, and service, go to the EMC Powerlink

website (registration required) at http://Powerlink.EMC.com.

Troubleshooting – For troubleshooting information, go to Powerlink, search for Celerra Tools, and select Celerra Troubleshooting from the navigation panel on the left.

Technical support – For technical support, go to Powerlink and choose Support. On the Support page, you can access Support Forums, request a product enhancement, talk directly to an EMC representative, or open a service request. To open a service request, you must have a valid support agreement. Please contact you EMC sales representative for details about obtaining a valid support agreement or to answer any questions about your account.

Note: Do not request a specific support representative unless one has already been assigned to your particular system problem.

Problem Resolution Roadmap for EMC Celerra contains additional information about using Powerlink and resolving problems.

E-Lab Interoperability Navigator

The EMC E-Lab Interoperability Navigator is a searchable, web-based application that provides access to EMC interoperability support matrices. It is available on the EMC Powerlink website at http://Powerlink.EMC.com. After logging in to Powerlink, go to Support > Interoperability and Product Lifecycle Information > E-Lab Interoperability Navigator.

Log files for troubleshooting

The following log files are available to help troubleshoot replication:

◆ server_log for messages

◆ /nas/log/sys_log for messages

◆ /nas/log/cmd_log for internal commands run for replication

Additionally, use the fs_replicate –info –verbose <number_of_lines> command to view the delta-set status.

Page 155: Celerra_Replcatr

155 of 172Version 5.6.47Using Celerra Replicator (V1)

server_log messages

Table 13 on page 155 shows an example message generated in the server_log for a Data Mover performing replication at the source site:

Replication::Valid v:CISn363 Delta:2564 ad:525312 g:2564 nc:1

Table 14 on page 155 shows an example message generated in the server_log for a Data Mover performing replication at the destination site:

Playback: v:361, Delta:3557, g:3557, ad:263168, nDelta:7

Table 13 Sample server_log message from source site

v: Volume ID of the source file system. In the above example, 363 is the volume ID.

Delta: Delta-set number copied to the SavVol.

ad: Address at the block level (1 block = 512 bytes) on the SavVol where the delta set is created.

g: Chunk number on the SavVol. If all delta sets consist of one chunk, this number is the same as the delta-set number.

nc: Number of chunks in the delta set. One chunk equals 128 MB.

Table 14 Sample server_log message from destination site

v: Volume ID of the destination file system.

Delta: First delta-set number in the group replayed to the destination file system.

g: First chunk number in the first delta set. If all the delta sets consist of one chunk, this number is the same as the delta-set number.

ad: Address on the SavVol where the first delta set in the group is located.

nDelta: Number of delta sets in the group replicated to the destination file system.

Page 156: Celerra_Replcatr

Using Celerra Replicator (V1)156 of 172 Version 5.6.47

Network performance troubleshooting

When you experience a performance issue during the transferring of delta sets, check the following:

◆ Duplex network configuration mismatch (Full, Half, Auto, and so on) between the Data Mover and the network switch

◆ Packet errors and input/output bytes using the server_netstat –i command on the Data Mover

◆ Packet errors and input/output bytes on the network switch port

◆ Transfer rate in the fs_replicate -info command output

Create a file from the network client, refresh the source file system, and then check the transfer rate in the fs_replicate -info command output.

Failure during transport of delta set

If a network failure occurs during the transport of a delta set over IP, the replication service continues to resend this delta set until it is accepted by the destination site.

Failure of fs_copy command process

If the output of the fs_copy -list command reports failed in the Status field, or the /nas/log/sys_log has one or more Copy failed messages, then the fs_copy process has failed and you must do the following:

1. Abort the fs_copy operation by running the command fs_copy -abort.

2. Start replication from scratch, as described in "Start replication from scratch" on page 129.

The following /nas/log/sys_log excerpt shows messages related to an fs_copy failure:

Aug 29 23:00:38 2006 VMCAST:3:9 Slot 2: 1156910199: Group:4673_0001854002020039

FSID:4666 Resync Copy failed. Full Copy has to be started. (t:17978065009336)

Aug 29 23:00:38 2006 VMCAST:3:2 Slot 2: 1156910199: Group:4673_0001854002020039

FSID:4666 Copy Failed restartad:0x0 (t:17978065072171)

Control Station restarts during replication

During replication processing:

◆ If the Control Station restarts during an fs_copy, you must abort the file system copy using the fs_copy -abort command for the source and destination file systems. Then start the copy again.

◆ If the Control Station restarts with active file system replications, the replication sessions are not affected and continue to run.

Page 157: Celerra_Replcatr

157 of 172Version 5.6.47Using Celerra Replicator (V1)

Control Station fails over

For remote replication, if a primary Control Station fails over to a standby Control Station, the replication service continues to run but the replication management capabilities are unavailable. For example, you cannot perform list, abort, start, or refresh replication functions from the standby Control Station. First resolve the problem with the primary Control Station before executing any of these commands. This applies to the Control Stations at the source and destination sites.

NS series loses power

If any of the NS series systems or a CNS-14 or NSX series system attached to a CLARiiON storage system loses power, after you power up the system, determine if you can restart replication using an incremental copy of the source file system as described in "Restarting a replication relationship" on page 89. If not, abort as detailed in "Abort Celerra Replicator" on page 79 and start replication from the beginning as described in "Initiating replication" on page 39.

Return codes for fs_copy

This section describes the possible fs_copy return codes that can be helpful when error checking scripts using the fs_copy command.

Table 15 on page 157 lists the possible return codes, and Table 16 on page 158 lists the return codes with corresponding error message IDs, a brief description, and error severity. To obtain detailed information about a particular error, use the nas_message -info <error_id> command.

Table 15 Return codes and description

Return code Description

0 Command completed successfully.

1 CLI usage error.

2 The object ID specified in the command line does not exist or is invalid.

3 Unable to acquire locks or resources for the command to start successfully.

4 Not applicable.

5 Possible communication error.

6 Transaction error reported during internal process transaction or during object verification or post transaction processing.

7 Data Mover error.

8 Not applicable.

Page 158: Celerra_Replcatr

Using Celerra Replicator (V1)158 of 172 Version 5.6.47

Table 16 fs_copy return codes with error IDs and message descriptions (page 1 of 3)

Return code

Error ID Message description

0 10246 This warning indicates that the time skew between the local and remote Celerra may have exceeded 10 minutes or that there is a passphrase mismatch between the local and remote Celerra.

0 10247 This warning indicates that the remote Celerra could not be reached with an http connection.

1 2102 This CLI usage error indicates that an invalid command option, such as the file system name, Celerra system name, interface name, or IP address, was specified.

1 10142 This CLI usage error indicates that an invalid interface or IP address was specified for the local or remote Celerra system.

2 2211 This error occurred because an invalid ID was specified.

3 2201 This error occurred because a process is not able to lock or reserve a resource, such as a file system, Data Mover, or interface.

5 4001 This is a communication error that occurred because the Data Mover is not connected.

5 4002 This communication error occurred because either the source Data Mover or the destination Data Mover (on which destination file system is mounted) is not connected.

5 4036 This communication error indicates that the interface specified for the source Data Mover is not connected to the interface on the destination Data Mover.

6 2001 This error occurred during the transaction when another process was modifying the database being accessed by the current command.

6 2103 This CLI usage error occurred during argument verification and indicates that an invalid command option, such as convert, autofullcopy, qos, or resync, was specified.

6 2207 This error occurred when the transaction was aborted due to an internal failure or abnormal termination of the command.

6 2225 This error indicates that an invalid checkpoint was specified in the -fromfs option of the differential copy argument or that the abort command was issued and the file system is involved in multiple copy sessions. The copy session should be aborted using the destination file system.

Page 159: Celerra_Replcatr

159 of 172Version 5.6.47Using Celerra Replicator (V1)

6 2227 This error indicates that the specified destination file system has FLR status of Enterprise. FLR-C enabled file systems (source or destination) are not supported for fs_copy or Celerra Replicator (V1).

6 2237 This error occurred during the transaction as specified in the error message.

6 2241 This error indicates that the command executed on the remote system failed.

6 2243 This error indicates that the checkpoint specified in the command is inactive.

6 2245 This error indicates that the file system type is invalid.

6 3105 This error indicates that either 1) an invalid file system is specified for the copy session or the copy session already exists, 2) the file system is not part of the specified session or, 3) the restart checkpoint specified is not newer than the destination file system.

6 3128 This error indicates that the source file system was restored from the checkpoint.

6 3134 This error indicates that the checkpoint specified is not newer

than the destination file system.

6 3136 This error occurred because the destination file system has replication set up.

6 3138 This error indicates that the file system is already part of another copy session.

6 3139 This error indicates that the file system is not part of a copy session.

6 4019 This error occurred when polling the progress of the copy session and indicates that the command failed to complete.

6 4103 This error occurred because the file system is not mounted.

6 4109 This error occurred because the file system is not mounted.

6 4111 This error occurred because the file system is mounted read/write by the server specified in the message.

6 4124 This error occurred because the file system is mounted read/write by the server specified in the error.

Table 16 fs_copy return codes with error IDs and message descriptions (page 2 of 3)

Return code

Error ID Message description

Page 160: Celerra_Replcatr

Using Celerra Replicator (V1)160 of 172 Version 5.6.47

6 4205 This error occurred because there is an invalid interface specified for the source file system. Check the interfaces specified for the Data Mover on which the file system is mounted.

6 4420 This error occurred because the destination file system is a backup of another replication or copy session.

6 4424 This error occurred because fsck is being executed on the file system.

6 4425 This error occurred because aclck is being executed on the file system.

6 4446 This error occurred because the remote Celerra system is running an out-of-family NAS code version with respect to the NAS code version running on the local Celerra system.

6 4447 This error occurred because IP Alias is configured on the system and the remote system's version could not be identified.

6 5000 This error occurred during the transaction because of an internal error.

6 10233 This error indicates that the query executed on the remote system failed.

6 10272 This error occurred because the destination file system could not found.

6 10273 This error indicates that the copy session failed.

6 10274 This error occurred because either the user aborted the copy session or there is a problem with the Data Mover.

6 10277 This error indicates that the checkpoint specified is older than than the replication configured on the source file system.

6 10299 This error occurred because a writable checkpoint is specified in the CLI.

6 10311 This error indicates that the clean up process for the copy session on destination failed.

6 10312 This error occurred because the size of the source and destination file systems does not match.

7 10310 This error occurred because the system was unable to retrieve copy session information from the Data Mover.

Table 16 fs_copy return codes with error IDs and message descriptions (page 3 of 3)

Return code

Error ID Message description

Page 161: Celerra_Replcatr

161 of 172Version 5.6.47Using Celerra Replicator (V1)

Error messages for Celerra Replicator

As of version 5.6, all new event, alert, and status messages provide detailed information and recommended actions to help you troubleshoot the situation.

To view message details, use any of the following methods:

◆ Celerra Manager:

• Right-click an event, alert, or status message and select to view Event Details, Alert Details, or Status Details.

◆ Celerra CLI:

• Use the nas_message -info <error_id> command to retrieve the detailed information for a particular error.

◆ EMC Celerra Network Server Error Messages Guide:

• Use this guide to locate information about messages that are in the earlier-release message format.

◆ Powerlink:

• Use the text from the error message’s brief description or the message’s ID to search the Knowledgebase on Powerlink. After logging in to Powerlink, go to Support > Knowledgebase Search > Support Solutions Search.

Page 162: Celerra_Replcatr

Using Celerra Replicator (V1)162 of 172 Version 5.6.47

Related information

Specific information related to the features and functionality described in this document is included in:

◆ EMC Celerra Glossary

◆ EMC Celerra Network Server Command Reference Manual

◆ EMC Celerra Network Server Error Messages Guide

◆ EMC Celerra Network Server Parameters Guide

◆ EMC Celerra Network Server Release Notes

◆ Configuring EMC Celerra Events and Notifications

◆ Configuring EMC Celerra Time Services

◆ Managing EMC Celerra Volumes and File Systems Manually

◆ Managing EMC Celerra Volumes and File Systems with Automatic Volume Management

◆ Online Celerra man pages

◆ Problem Resolution Roadmap for EMC Celerra

◆ Replicating EMC Celerra CIFS Environments (V1)

◆ Using EMC Celerra FileMover

◆ Using EMC Celerra Replicator for iSCSI (V1)

◆ Using File-Level Retention on EMC Celerra

◆ Using International Character Sets with EMC Celerra

◆ Using SnapSure on EMC Celerra

◆ Using TimeFinder/FS, NearCopy, and FarCopy with EMC Celerra

The EMC Celerra Network Server Documentation CD, supplied with Celerra and also available on the EMC Powerlink website, provides the complete set of EMC Celerra customer publications. After logging in to Powerlink, go to Support > Technical Documentation and Advisories > Hardware/Platforms Documentation > Celerra Network Server. On this page, click Add to Favorites. The Favorites section on your Powerlink home page provides a link that takes you directly to this page.

Celerra Support Demos are available on Powerlink. Use these instructional videos to learn how to perform a variety of Celerra configuration and management tasks. After logging in to Powerlink, go to Support > Product and Diagnostic Tools > Celerra Tools > Celerra Support Demos.

Page 163: Celerra_Replcatr

163 of 172Version 5.6.47Using Celerra Replicator (V1)

Training and Professional Services

EMC Customer Education courses help you learn how EMC storage products work together within your environment in order to maximize your entire infrastructure investment. EMC Customer Education features online and hands-on training in state-of-the-art labs conveniently located throughout the world. EMC customer training courses are developed and delivered by EMC experts. Go to EMC Powerlink at http://Powerlink.EMC.com for course and registration information.

EMC Professional Services can help you implement your Celerra Network Server efficiently. Consultants evaluate your business, IT processes, and technology and recommend ways you can leverage your information for the most benefit. From business plan to implementation, you get the experience and expertise you need, without straining your IT staff or hiring and training new personnel. Contact your EMC representative for more information.

Page 164: Celerra_Replcatr

Using Celerra Replicator (V1)164 of 172 Version 5.6.47

Appendix A: fs_replicate -info output fields

The first part of the output from the fs_replicate -info command presents information about the source file system. Table 17 on page 164 describes these output fields.

Table 17 First section of fs_replicate -info output fields

Field Description

id Source file system ID.

name Source file system name.

fs_state Condition of the source file system. Indicates whether the source file system is active (active), mounted read-only (romounted), or unmounted (frozen).

type Whether it is a replication or playback service.

replicator_state Indicates if the replication service is active, inactive, or creating a delta set. If inactive, replication has fallen out-of-sync.

source_policy How replication handles flow-control situations. Values are ReadOnly, Freeze, or NoPolicy, where:

ReadOnly = the file system is only available for reads.

Freeze = reads and writes are not allowed.

NoPolicy = no policy is defined.

high_water_mark Size in MB of the file system changes accumulated since the last delta set was created. When this size is reached, the replication service automatically creates a delta set on the SavVol.

time_out Interval (in seconds) when the replication service creates a delta set.

current_delta_set Current delta set being processed. This reflects the current delta set being tracked in memory.

current_number_of_blocks

Number of modified blocks in the current delta set. One block size is 8 KB.

flow_control Indicates if there is enough space in the SavVol to write a new delta set to the SavVol.

Note: An active status indicates there is not enough SavVol space to process the next delta set.

total_savevol_space The total size of the SavVol.

savevol_space_available The amount of free space in the SavVol.

Page 165: Celerra_Replcatr

165 of 172Version 5.6.47Using Celerra Replicator (V1)

The second part of the output from the fs_replicate command presents information about the destination file system. Table 18 on page 165 describes these output fields.

Table 18 Second section of fs_replicate -info output fields

Field Description

id Destination file system id.

name Destination file system name.

type Indicates whether it is a replication or playback service.

playback_state Indicates whether the playback service is active, inactive, or replaying a delta set. If inactive, replication has not started.

high_water_mark Size in MB of the delta sets at which the replication service automatically replays a delta set on the SavVol.

time_out Interval (in seconds) when the replication service creates or replays a delta set.

current_delta_set Lists the next delta set to replay to the destination file system.

flow_control Indicates whether there is enough space to write an incoming delta set to the destination SavVol.

Note: An active status indicates there is not enough SavVol space to process the next delta set.

total_savevol_space The total size of the SavVol.

savevol_space_available The amount of free space in the SavVol.

outstanding delta sets The delta sets on the destination SavVol the replication service has not played back to the destination file system.

Page 166: Celerra_Replcatr

Using Celerra Replicator (V1)166 of 172 Version 5.6.47

The third part of the output from the fs_replicate command presents information on the communication state between the source and destination Data Movers. Table 19 on page 166 describes these output fields.

When using the verbose option with the fs_replicate -info command, the replication service generates this additional output. Table 20 describes these output fields.

Table 19 Third section of fs_replicate -info output fields

Field Description

communication_state Indicates whether the Data Movers on each Celerra system can communicate with each other.

current_transfer_rate The rate at which the last 128 MB of data was sent across the IP network.

avg_transfer_rate The average rate at which the last 128 sets of data were sent across the IP network.

source_ip IP address of the source Data Mover.

source_port Port number of the source Data Mover. This is assigned dynamically.

destination_ip IP address of the destination Data Mover.

destination_port Port number of the destination Data Mover. This is always assigned the number 8888 by default.

QOS_bandwidth The IP network bandwidth throttle used for this replication relationship.

Zero (0) represents maximum available network bandwidth.

Table 20 fs_replicate -info -verbose output fields (page 1 of 2)

Field Description

Source file system

Delta Delta-set number or delta-set ID that was created.

Create Time Date and start time the delta set was created.

Dur Duration of time (in seconds) to create the delta set.

Blocks Number of blocks modified in the delta set.

Page 167: Celerra_Replcatr

167 of 172Version 5.6.47Using Celerra Replicator (V1)

Destination file system

Playback Time Date and start time the delta set replayed to the destination file system.

Dur Duration of time (in seconds) to replay the delta set or DsinGroup.

Blocks Number of blocks modified in the delta set.

DSinGroup Number of delta sets in the group played back. In some instances, the playback service can play back more than one delta set at once. In this case, the Dur and Blocks fields refer to the group as a whole, not an individual delta set.

Table 20 fs_replicate -info -verbose output fields (page 2 of 2)

Field Description

Page 168: Celerra_Replcatr

Using Celerra Replicator (V1)168 of 172 Version 5.6.47

Page 169: Celerra_Replcatr

169 of 172Version 5.6.47Using Celerra Replicator (V1)

Index

Aaborting replication 79Automatic File System Extension 98, 100, 104

definition 5extending after replication failure 104recovering from failure 100starting replication from scratch 104

Bbandwidth size

changing policy 109modifying 109

Ccautions

graceful shutdown 9serial replication sessions 9system 9Unicode to ASCII replication 9

Celerra Replicatorcautions 9checking status of 55, 114log files 154restarting replication 89restrictions 7starting replication 48system requirements 22upgrade considerations 24, 25

checkpointdefinition 6

Commandsfs_replicate, flow control options 32nas_fs, calculate SavVol size 33

commandsfs_ckpt, copy checkpoints 50fs_ckpt, creating 43fs_ckpt, using 59fs_copy, checkpoint copy 46fs_copy, copy checkpoints 45fs_copy, copy incremental changes 52fs_copy, events for 122fs_copy, using 46, 52fs_replicate, aborting 79fs_replicate, changing bandwidth 109fs_replicate, check status 55fs_replicate, control SavVol size 34fs_replicate, description of output fields 165fs_replicate, failover 64fs_replicate, failure 50fs_replicate, output definitions 165fs_replicate, resynchronize 68fs_replicate, reverse 76fs_replicate, starting 48fs_replicate, suspending 81fs_replicate, using 57

nas_cel, verify Control Station relationship 18, 41nas_cel, view passphrase 42nas_fs, creating destination 45server_mount, mounting file systems 45server_mountpoint, creating 45server_sysstat, monitor memory usage 35

configurationconsiderations 35replication policies 30

Control Stationpreinitialization 40

Ddata flow control 31delta set

checking status 115definition 6minimum size 30overview 30

delta-settransport failure 156

disk transport 140

Eexit codes for fs_copy 157extending file system

after replication failover 104

Ffailover

initiate 64options 62

failuredelta-set transport 156

failure, fs_replicate 50file system

automatically extending size 98events 121manually extending size 101

flow controlfreeze 107read-only 107

fs_copy return codes 157

Hhigh water mark

definition 6resetting policy 105

HTTP communication 17

Iinformation, related 163Initial Copy

transporting by disk 144initial copy

transporting by disk 140transporting by tape 144

Page 170: Celerra_Replcatr

170 of 172 Using Celerra Replicator (V1)Version 5.6.47

IP replication servicedefinition 6

Llocal replication

definition 6description of 10process overview 10system requirements 22

log files 154loopback replication, definition 6losing power, CLARiiON storage system 157

Mmessages, server_log 155

Nnas_cel, using 41nas_fs, using 45, 101network performance, troubleshooting 156NS series, loses power 157

Ooverview

local replication 10remote replication 12replication failover 13replication reverse 17suspend 81

Pparameters, changing Replicator SavVol size 123passphrase

establishing 17viewing 42

physical transport of replication data 139playback service

checking 115definition 6

policybandwidth size 109flow control 31high watermark 30time-out 30

Rrecovering from a corrupted file system using nas_fsck 130recovering from auto extension failure 100remote replication

copycheckpoint 45incremental changes 52

createcheckpoint 42destination file system 45

definition 6

description of 11establish communication between Celerra systems 17process overview 12start replication 48system requirements 22verify communication between Celerra systems 41

replicationaborting 79starting from scratch 104

replication failoverdefinition 6process overview 13

replication policyconfiguring 35flow control 107high watermark 30setting 105time-out 30triggers 35

replication service, definition 6replications per Data Mover 34Replicator SavVol

definition 6parameter for changing size 123size requirements 33

restartout of synchronization

checkpoints 21, 59suspended replication 20

restrictionssystem 7TimeFinder/FS 7TimeFinder/FS Near and Far Copy 7

resynchronizeprocess overview 16

reversing replicationafter failover 75definition 6maintenance procedure 111process overview 17

Sserver_df, monitoring file systems 114server_log 155server_param command, SavVol default 123server_sysstat, monitoring replication 114SnapSure, SavVol definition 6SNMP traps 121status

checking 55of replication 114

suspendoverview 81procedure 20

systemrequirements 22restrictions 7

Page 171: Celerra_Replcatr

171 of 172Version 5.6.47Using Celerra Replicator (V1)

Ttape, transporting initial copy procedure 144timeout

definition 6replication policy 30resetting policy 105

triggers, data flow 31troubleshooting

accommodating network concerns 127calculating modifications rate on source file system 127changing the passphrase 124Control Station failovers 157Control Station reboots during replication 156copying a file system to multiple destinations with fs_copy 136creating checkpoints on the destination site 136creating restartable checkpoints 125enlarging SavVol size 126error messages 162failure during transport of delta set 156network performance 156NS series loses power 157reading server_log messages 155recovering from an inactive replication state 135return codes for fs_copy 157starting replication from a differential copy 128starting replication from scratch 129temporarily mount destination file system read/write 133using log files 154

trust relationship 17

Uupgrading 24, 25

VVirtual Data Mover, definition 7virtual provisioning

definition 7

Page 172: Celerra_Replcatr

About this document

As part of its effort to continuously improve and enhance the performance and capabilities of the Celerra Network Server product line, EMC periodically releases new versions of Celerra hardware and software. Therefore, some functions described in this document may not be supported by all versions of Celerra software or hardware presently in use. For the most up-to-date information on product features, see your product release notes. If your Celerra system does not offer a function described in this document, contact your EMC Customer Support Representative for a hardware upgrade or software update.

Comments and suggestions about documentation

Your suggestions will help us improve the accuracy, organization, and overall quality of the user documentation. Send a message to [email protected] with your opinions of this document.

Copyright © 1998-2009 EMC Corporation. All rights reserved.

EMC believes the information in this publication is accurate as of its publication date. The information is subject to change without notice.

THE INFORMATION IN THIS PUBLICATION IS PROVIDED “AS IS.” EMC CORPORATION MAKES NO REPRESENTATIONS OR WARRANTIES OF ANY KIND WITH RESPECT TO THE INFORMATION IN THIS PUBLICATION, AND SPECIFICALLY DISCLAIMS IMPLIED WARRANTIES OF MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE.

Use, copying, and distribution of any EMC software described in this publication requires an applicable software license.

For the most up-to-date regulatory document for your product line, go to the Technical Documentation and Advisories section on EMC Powerlink.

For the most up-to-date listing of EMC product names, see EMC Corporation Trademarks on EMC.com.

All other trademarks used herein are the property of their respective owners.

Version 5.6.47 172 of 172