recover point implementation lab guidev3.2
TRANSCRIPT
RecoverPoint Implementation
September 2009
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 2 of 223
Copyright
Copyright © 1996, 2000, 2001, 2002, 2003, 2004, 2005, 2006, 2007, 2008, 2009 EMC Corporation. All Rights Reserved. EMC believes the information in this publication is accurate as of its publication date. The information is subject to change without notice. THE INFORMATION IN THIS PUBLICATION IS PROVIDED “AS IS.” EMC CORPORATION MAKES NO REPRESENTATIONS OR WARRANTIES OF ANY KIND WITH RESPECT TO THE INFORMATION IN THIS PUBLICATION, AND SPECIFICALLY DISCLAIMS IMPLIED WARRANTIES OF MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Use, copying, and distribution of any EMC software described in this publication requires an applicable software license. EMC, ICDA® (Integrated Cached Disk Array), and EMC2® (the EMC logo), and Symmetrix®, are registered trademarks of EMC Corporation. EMC™ and SRDF™ are trademarks of EMC Corporation. All other trademarks used herein are the property of their respective owners.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 3 of 223
Trademark Information
EMC Trademarks
EMC2, EMC, Symmetrix, Celerra, CLARiiON, CLARalert, Connectrix, Dantz, Documentum, HighRoad, Legato, Navisphere, PowerPath, ResourcePak, SnapView/IP, SRDF, TimeFinder, VisualSAN, and where information lives are registered trademarks and EMC Automated Networked Storage, EMC ControlCenter, EMC Developers Program, EMC OnCourse, EMC Proven, EMC Snap, Access Logix, AutoAdvice, Automated Resource Manager, AutoSwap, AVALONidm, C-Clip, Celerra Replicator, Centera, CentraStar, CLARevent, CopyCross, CopyPoint, DatabaseXtender, Direct Matrix, Direct Matrix Architecture, EDM, E-Lab, Enginuity, FarPoint, FLARE, GeoSpan, InfoMover, MirrorView, NetWin, OnAlert, OpenScale, Powerlink, PowerVolume, RepliCare, SafeLine, SAN Architect, SAN Copy, SAN Manager, SDMS, SnapSure, SnapView, StorageScope, SupportMate, SymmAPI, SymmEnabler, Symmetrix DMX, Universal Data Tone, and VisualSRM are trademarks of EMC Corporation. All other trademarks used herein are the property of their respective owners.
Third Party Trademarks
AIX is a registered trademark of International Business Machines Corporation. Brocade, SilkWorm, SilkWorm Express, and the Brocade logo are trademarks or registered trademarks of Brocade Communications Systems, Inc., in the United States and/or in other countries. Compaq and the names of Compaq products referenced herein are either trademarks and/or service marks or registered trademarks and/or service marks of Compaq. Hewlett-Packard, HP, HP-UX, OpenView, and OmniBack are trademarks, or registered trademarks of Hewlett-Packard Company. McDATA, the McDATA logo, and ES-2500 are registered trademarks of McDATA Corporation. Microsoft, Windows, and Windows NT are either registered trademarks or trademarks of Microsoft Corporation in the United States and/or other countries. NobleNet is a registered trademark of Rogue Wave Software, Inc. SANbox is a trademark of QLogic Corporation. Sun, Sun Microsystems, the Sun Logo, SunOS and all Sun-based trademarks and logos, Java, the Java Coffee Cup Logo, and all Java-based trademarks and logos, Solaris, and NFS, are trademarks or registered trademarks of Sun Microsystems, Inc. in the United States and other countries. UNIX is a registered trademark of The Open Group.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 4 of 223
Document Revision History
Rev # File Name Date 1.0 Deployment Workshop 3.0.doc February, 2008 1.1 Deployment Lab Guide 3.0.doc April, 2008 2.0 RP3.1 Workshop Lab Guide.doc December, 2008 2.1 RP3.1.1_Workshop_Labguide.doc July, 2009 2.2 RPImp_Labguide.doc September, 2009
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 5 of 223
Table of Contents
Lab Exercise 1: SAN Configuration for RecoverPoint ......................................................7
Part 1: Set initial System Layout and Discover WWN’s ......................................................8 Part 2: Zoning the RPAs in an MDS Series SAN..............................................................12 Part 3: Zoning the RPAs in a Brocade SAN......................................................................19 Part 4: LUN Masking for the RPA – Symmetrix ................................................................25 Part 5: LUN Masking for the RPA – CLARiiON.................................................................29
Lab Exercise 2: Installation of RecoverPoint Appliances...............................................36 Part 1: RPA Cluster Installation ........................................................................................37 Part 2: RPA Installation – Add RecoverPoint License.......................................................51
Lab Exercise 3: Installing Host Splitter Drivers...............................................................53 Part 1: Host Splitter Driver Installation - Windows ............................................................54 Part 2: Host Splitter Driver Installation – Solaris ...............................................................58 Part 3: Host Splitter Driver Installation – AIX ....................................................................62 Part 4: Add Splitters Using the RecoverPoint Management Application ...........................66
Lab Exercise 4: Replication Configuration ......................................................................68 Part 1: Create Consistency Group....................................................................................69
Lab Exercise 5: Managing Replication Jobs....................................................................78 Part 1: CRR Image Access...............................................................................................79 Part 2: CRR Failover/Failback ..........................................................................................87 Part 3: CRR Recovery Production ....................................................................................95
Lab Exercise 6: Working with Consistency Group Policies .........................................101 Part 1: Consistency Group Policies ................................................................................102
Lab Exercise 7: Install and Configure CLARiiON Array Based Splitter.......................112 Part 1: Software tasks on the CX array...........................................................................113 Part 2: Provisioning RPAs ..............................................................................................118 Part 3: Add CLARiiON Splitter to the RecoverPoint Cluster ...........................................121
Lab Exercise 8: Migrate to Cisco SANTap Fabric Splitter ............................................125 Part 1: Prepare to Migrate to SANTap............................................................................126 Part 2: Deploying SANTap..............................................................................................131 Part 3: Add RecoverPoint to the SANTap Environment..................................................143 Part 4: Restart Replication: SANTap Implementation Complete for all Switches............151
Lab Exercise 9: Migrate to Brocade Fabric Splitter ......................................................152 Part 1: Prepare to Migrate to Brocade SAS....................................................................153 Part 2: Verify AP-7600B is prepared for RecoverPoint Installation .................................156 Part 3: Verify the SysVI exists and check fabric logins ...................................................161 Part 4: Configure Zoning – Frame Redirect ....................................................................163
Lab Exercise 10: Working With CLR...............................................................................169 Part 1: Add CDP Copy to CRR Consistency Group........................................................170 Part 2: CDP Failover.......................................................................................................175 Part 3: CDP Restore .......................................................................................................180 Part 4: Perform CRR Failover with a CLR Consistency Group .......................................186
Lab Exercise 11: Upgrade RPAs.....................................................................................192
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 6 of 223
Part 1: Prepare Consistency Groups ..............................................................................193 Part 2: Download ISO Image ..........................................................................................195 Part 3: Reapply Settings .................................................................................................198 Part 4: Working with RecoverPoint 3.2 NDU Tool...........................................................202
Lab Exercise 12: Troubleshooting..................................................................................205 Part 1: Saving and Restoring RPA Settings....................................................................206 Part 2: Implementing SNMP Traps / Email Alert / Syslog Reporting...............................208 Part 3: Troubleshooting Using the RecoverPoint Management Application....................210 Part 4: Running Log Collection .......................................................................................213 Part 5: Performance and Statistics .................................................................................218
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 7 of 223
Lab Exercise 1: SAN Configuration for RecoverPoint
Purpose: To configure an EMC SAN environment to function with RecoverPoint appliances.
Tasks: In this lab, you perform the following tasks: • Identify WWNs of the RPAs
• Identify storage and host WWNs
• Identify LUNs to be used for repository, journal and host data (replica) volumes
• Implement zoning and LUN masking to support the RPA
References: • Installation Guide • Administrator’s Guide • Lecture material
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 8 of 223
Part 1: Set initial System Layout and Discover WWN’s Step Action
Note: If using Windows host, download Putty.exe or any other utility that supports SSH.
1 Connect to the first RPA at local site via SSH. Connect into the management IP interface. Use the default user boxmgmt and password boxmgmt to log in.
2 For the sake of lab activities the RPAs management IPs have been defined prior to the class so everything can be performed remotely from the classroom. During this step do not configure RPA management IP address
Note: Perform these steps in all RPAs at each site.
login as: boxmgmt Using keyboard-interactive authentication. Password: Linux RecoverPoint_s1-bos1_RPA2 2.6.27.6-k100.x64.deb #24 SMP Mon Apr 6 01:27:58 EDT 2009 x86_64 Loading RPA settings... done Initializing Installation Manager... done Installation Manager - RecoverPoint Version 3.2(e.146) Do you want to configure a temporary IP address? (y/n)? n ** Main Menu ** [1] Installation [2] Setup [3] Diagnostics [4] Cluster operations [5] Shutdown / Reboot operations [Q] Quit
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 9 of 223
Step Action 3 From the boxmgmt Main menu select Diagnostics > Fibre Channel diagnostics > View
Fibre Channel details.
Fibre Channel Details
Port 0 ------------------------------------- |----------------|--------------------| | | | |----------------|--------------------| | Link status | Up | | wwn | 0x50012482001063a0 | | node wwn | 0x50012482001063a1 | | port id | 4f0000 | | operating mode | point to point | | speed | 2 GB | |----------------|--------------------| Port 1 ------------------------------------- |----------------|--------------------| | | | |----------------|--------------------| | Link status | Up | | wwn | 0x50012482013063a0 | | node wwn | 0x50012482013063a1 | | port id | ce0005 | | operating mode | point to point | | speed | 2 GB |
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 10 of 223
Step Action 4 Local Site. Record the Port WWN for each FC Port below.
Site 1 RPA 1: Port 0 _______________________________________ Port 1 _______________________________________ Port 2 _______________________________________ Port 3 _______________________________________ Site 1 RPA 2: Port 0 _______________________________________ Port 1 _______________________________________ Port 2 _______________________________________ Port 3 _______________________________________
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 11 of 223
Step Action 5 Remote Site. Record the Port WWN for each FC Port below.
Site 2 RPA 1: Port 0 _______________________________________ Port 1 _______________________________________ Port 2 _______________________________________ Port 3 _______________________________________ Site 2 RPA 2: Port 0 _______________________________________ Port 1 _______________________________________ Port 2 _______________________________________ Port 3 _______________________________________
End of Lab Exercise
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 12 of 223
Part 2: Zoning the RPAs in an MDS Series SAN Step Action
Before we can replicate volumes using RecoverPoint, we must first expose RPAs HBAs to host HBAs, as well as to storage ports involved in replication at each site. This means a lot of zoning. At the end of this exercise, in the zones database of every switch there should be several zones representing every component involved in the replication scenario, for example:
a) RPAs to first Windows host b) RPAs to second Windows host c) RPAs to AIX d) RPAs to ESX e) RPAs to CLARiiON SPA port f) RPAs to CLARiiON SPB port g) RPAs to Symmetrix port
Each group will work in a different environment with different switches, hosts and storage ports so there should not be conflicts between groups. Refer to the lab environment reference document for your specific components. Note: This lab exercise shows the steps to create zones using command line. If you feel more comfortable working with the GUI, verify with your instructor if there is a host where Fabric Manager is installed. If so, feel free to use it. Just make sure you create all required zones as showed here.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 13 of 223
Step Action 1 Telnet as admin on your first switch at local site.
Verify which RPA ports are logged into this switch by viewing the flogi database. If you have two RPAs per site and two fabrics at each site, each fabric will have four RPA ports. RPAs WWN start with 50:01:24:82:xx:xx:xx:xx. Write down the RPA WWNs connected to this switch (port name column).
Command Syntax: show flogi database vsan <vsan#> Command Example: mds# show flogi database vsan 20
Command Output: show flogi database without vsan #
Note that the VSAN number may be different in your assigned switches. List the existent VSANs by running “show vsan”. Also, there should only be one production VSAN with operational state up and an active zoneset (which will be the VSAN to use). Verify that by running “show zoneset active”. Important: Configuration changes on the switch should only be performed by one user at a time. Each workgroup should coordinate switch configurations to ensure only one user is changing a configuration. If for some reason RPA ports are in a different VSAN than hosts and storage ports, go ahead and move them to the same VSAN.
s1-mdsa-14-80# show flogi database -------------------------------------------------------------------------------- INTERFACE VSAN FCID PORT NAME NODE NAME -------------------------------------------------------------------------------- fc1/4 13 0xce0000 50:06:04:8a:d5:f0:f8:cf 50:06:04:8a:d5:f0:f8:cf fc1/8 30 0x1f0007 50:06:04:8a:d5:f0:f8:c1 50:06:04:8a:d5:f0:f8:c1 fc1/12 30 0x1f02ef 50:06:01:63:3c:e0:2e:6f 50:06:01:60:bc:e0:2e:6f fc1/13 30 0x1f03ef 50:06:01:68:3c:e0:2e:6f 50:06:01:60:bc:e0:2e:6f fc1/14 13 0xce0001 50:06:04:8a:d5:f0:f8:df 50:06:04:8a:d5:f0:f8:df fc1/15 20 0xbc0007 50:06:04:8a:d5:f0:f8:d0 50:06:04:8a:d5:f0:f8:d0 fc2/1 13 0xce0002 50:01:24:82:01:30:a7:a4 50:01:24:82:01:30:a7:a5 fc2/2 13 0xce0003 50:01:24:82:00:10:25:a5 50:01:24:82:00:10:25:a4 fc2/3 30 0x1f0003 50:01:24:82:01:30:b2:9f 50:01:24:82:01:30:b2:9e fc2/4 20 0xbc0005 50:01:24:82:01:30:dd:9f 50:01:24:82:01:30:dd:9e fc2/5 13 0xce0004 50:01:24:82:00:10:aa:a4 50:01:24:82:00:10:aa:a5 fc2/6 20 0xbc0006 50:01:24:82:00:10:c5:a3 50:01:24:82:00:10:c5:a2 fc2/7 20 0xbc0008 10:00:00:00:c9:5e:af:5d 20:00:00:00:c9:5e:af:5d fc2/8 1 0x340019 10:00:00:00:c9:37:0c:0a 20:00:00:00:c9:37:0c:0a
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 14 of 223
Step Action 2 Enter the configuration terminal.
Command Example: mds# conf t
3 TARGET Zones. Create a new zone. As per EMC Best Practices, you will add only two members to this zone. The first one is an RPA port connected to this fabric and the second one the host HBA.
Command Syntax: zone name <name> vsan <VSAN#>
Command Example: mds#(config) zone name rpa5_p0_win3a vsan 20
4 Add the pwwn of the first RPA port that is logged into this switch as member of the new zone. Use one of the WWNs saved on step 1.
Command Syntax: member pwwn <RPA PWWN> Command Example: mds(config-zone)# member pwwn 50:01:24:82:00:10:42:a5
5 Add the pwwn of the host HBA as a member of the new zone.
Command Syntax: member pwwn <HBA PWWN> Command Example: mds(config-zone)# member pwwn 10:00:00:00:c9:69:e1:34
There are many ways to find out the host HBA WWN. Choose the one you are more comfortable with. Hint, hosts already have zones created to the storage ports, looking at an existing zone may be the easiest way to discover that. Run “show zoneset active” to list the active zones in the database. But you can cross check that by looking direct at the host operating system.
6 You have created the first zone for your first host with one of the RPAs ports connected to this fabric. Exit the config-zone mode. This will return you to the config prompt.
Command Example: mds(config-zone)# exit
Now create additional zones using the same host HBA with all the other RPA ports in this
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 15 of 223
Step Action fabric. At the end, there should be four similar zones (you have created the first), with two members each, being one RPA port and the host HBA. Please repeat steps 3 through 6. Give different names to the new zones. Use the WWNs saved on step 1.
7 INITIATOR Zones. Create another new zone. As per EMC Best Practices, you will add only two members to this zone. The first one is an RPA port connected to this fabric and the second one the storage port.
Command Syntax: zone name <name> vsan <VSAN#> Command Example: mds#(config) zone name rpa5_p0_symm_fa4ba vsan 20
8 Add the pwwn of the first RPA port that is logged into this switch as member of the new
zone. Use one of the WWNs saved on step 1.
Command Syntax: member pwwn <RPA PWWN> Command Example: mds(config-zone)# member pwwn 50:01:24:82:00:10:42:a5
9 Add the pwwn of the storage port as a member of the new zone.
Command Syntax: member pwwn <Storage Port PWWN> Command Example: mds(config-zone)# member pwwn 50:06:04:8a:cc:d2:53:81
Verify which storage port to use in the lab environment reference document. There are many ways to find out the storage port WWN. Choose the one you are more comfortable with. Hint, storage ports already have zones created on this switch (zoned with hosts), looking at an existing zone may be the easiest way to discover that. Run “show zoneset active” to list the active zones in the database.
10 You have created the first zone for your first storage port with one of the RPAs ports connected to this fabric. If you are working in a CLARiiON environment you must create an additional zone to the second CLARiiON SP port with the same RPA port. Exit the config-zone mode. This will return you to the config prompt
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 16 of 223
Step Action
Command Example: mds(config-zone)# exit
Now create additional zones using the same storage port with all the other RPA ports in this fabric. At the end, there should be four (eight if CLARiiON in place) similar zones (you have created the first), with two members each, being one RPA port and the storage port. Please repeat steps 7 through 10. Give different names to the new zones. Use the WWNs saved on step 1.
11 Display the name of the active zoneset for your VSAN. You will need this name to add the new zones to the zoneset and activate the zoning changes in the next several steps. Add ‘do’ in front of the command if you are in the config prompt level.
Command Syntax: do show zoneset active vsan <VSAN#> Command Example: mds(config)# do show zoneset active vsan 20
Output: show zoneset active
12 From the config prompt add the new zones to the zoneset.
s1-mdsa-14-80# show zoneset active zoneset name bos1_rpa_BE_zoneset vsan 10 zone name s1_sun1_hba1_dmx155_fa16cA vsan 10 pwwn 10:00:00:00:c9:5e:af:d1 pwwn 50:06:04:8a:d5:f0:f8:cf zone name s1_rpa_all_dmx155_fa16cA vsan 10 pwwn 50:06:04:8a:d5:f0:f8:cf pwwn 50:01:24:82:01:30:a7:a4 pwwn 50:01:24:82:00:10:25:a5 pwwn 50:01:24:82:00:10:aa:a4 pwwn 50:01:24:82:01:30:63:a0
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 17 of 223
Step Action Command Syntax: conf t zoneset name <Zoneset Name> vsan # member <Zone Name> Command Example: mds# conf t mds(config)# zoneset name prod1_H2S_fabA vsan 20 mds(config-zoneset)# member rpa5_p0_win3a mds(config-zoneset)# member rpa5_p2_win3a mds(config-zoneset)# member rpa6_p0_win3a mds(config-zoneset)# member rpa6_p2_win3a mds(config-zoneset)# member rpa5_p0_symm_fa4ba mds(config-zoneset)# member rpa5_p2_symm_fa4ba mds(config-zoneset)# member rpa6_p0_symm_fa4ba mds(config-zoneset)# member rpa6_p2_symm_fa4ba
Note that this lab exercise only showed the steps to create zones from the RPAs to a host and zones from the RPAs to the storage port. You might need to create additional zones for the other hosts and storage ports. Refer to the lab environment reference document.
13 Exit the config-zoneset mode. This will return you to the config prompt.
Command Example: mds(config-zoneset)# exit
14 Activate the zoneset.
Command Syntax: zoneset activate name <Zoneset Name> vsan <VSAN#> Command Example: mds(config)# zoneset activate name prod1_H2S_fabA vsan 20
15 Exit the config mode. This will return you to the switch prompt.
Command Example: mds(config)# exit
16 Display the modified active zoneset for your VSAN. You should see your new zones there.
Command Syntax: show zoneset active vsan <VSAN#> Command Example: mds# show zoneset active vsan 20
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 18 of 223
Step Action
17 Copy the running configuration to the startup configuration so changes are persistent across reboots.
Command Syntax: mds# copy running-config startup-config Command Example: mds# copy run start
You have now taken care of zoning on the first fabric at local site. Telnet into your second
switch at local site and do the same thing (steps 1 to 17) – obviously the zone names and the zoneset name may be different there.
18 Check the zoning configuration from the RPAs. Log into each RPA as the boxmgmt user. From the Main menu select Diagnostics > Fibre Channel diagnostics > Detect Fibre Channel targets. Each ports should display the WWNs of the host HBA ports and storage ports.
Important: When you are done with zoning at local site (both switches), move on to remote
site. Telnet into your first and second switches at remote site and do the same thing (steps 1 to 18) – obviously the zone names and the zoneset name may be different there.
End of Lab Exercise
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 19 of 223
Part 3: Zoning the RPAs in a Brocade SAN Step Action
Before we can replicate volumes using RecoverPoint, we must first expose RPAs HBAs to host HBAs, as well as to storage ports involved in replication at each site. This means a lot of zoning. At the end of this exercise, in the zones database of every switch there should be several zones representing every component involved in the replication scenario, for example:
h) RPAs to first Windows host i) RPAs to second Windows host j) RPAs to AIX k) RPAs to ESX l) RPAs to CLARiiON SPA port m) RPAs to CLARiiON SPB port n) RPAs to Symmetrix port
Each group will work in a different environment with different switches, hosts and storage ports so there should not be conflicts between groups. Refer to the lab environment reference document for your specific components. Note: This lab exercise shows the steps to create zones using command line. If you feel more comfortable working with the GUI (WebTools) feel free to use it. Just make sure you create all required zones as showed here.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 20 of 223
Step Action 1 Telnet as admin on your first switch at local site (S1-brocade1_a, S1-brocade3_a etc).
Verify which RPA ports are logged into this switch. If you have two RPAs per site and two fabrics at each site, each fabric will have four RPA ports. The name server shows RPA entries as the ‘KASHYA’ vendor. RPAs WWN start with 50:01:24:82:xx:xx:xx:xx. Write down the RPA WWNs connected to this switch (port name).
Command Example: nsshow Output: nsshow
Important: Configuration changes on the switch should only be performed by one user at a time. Each workgroup should coordinate switch configurations to ensure only one user is changing a configuration.
DS_4100B:admin> nsshow { Type Pid COS PortName NodeName TTL(sec) N 010300; 3;50:01:24:82:00:10:4a:a0;50:01:24:82:00:10:4a:a1; na FC4s: FCP [KASHYA KBOX2_S2-BOS4 0001] Fabric Port Name: 20:03:00:05:1e:02:c6:63 Permanent Port Name: 50:01:24:82:00:10:4a:a0 Port Index: 3 Share Area: No Device Shared in Other AD: No Redirect: No N 010400; 3;50:06:04:8a:d5:f0:e6:40;50:06:04:8a:d5:f0:e6:40; na FC4s: FCP PortSymb: [38] "EMC SYMMETRIX 000190301081 SAF- 1cA " NodeSymb: [38] "EMC SYMMETRIX 000190301081 SAF- 1cA " Fabric Port Name: 20:04:00:05:1e:02:c6:63 Permanent Port Name: 50:06:04:8a:d5:f0:e6:40 Port Index: 4 Share Area: No Device Shared in Other AD: No Redirect: No N 010600; 3;50:01:24:82:01:30:e3:9d;50:01:24:82:01:30:e3:9c; na FC4s: FCP [KASHYA KBOX2_S2-BOS4 0001] Fabric Port Name: 20:06:00:05:1e:02:c6:63 Permanent Port Name: 50:01:24:82:01:30:e3:9d Port Index: 6 Share Area: No Device Shared in Other AD: No Redirect: No
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 21 of 223
Step Action 2 Display the effective configuration. You will need the configuration name to add new zones
to it.
Command Example: cfgactvshow
Output: cfgactvshow
DS_4100B:admin> cfgactvshow Effective configuration: cfg: L_Winspltr_allZONES_cfg zone: L_WINaf8f_15ca_1ca_SPa0 10:00:00:00:c9:5e:af:8f 50:06:01:60:3c:e0:09:c1 zone: L_WINspltr_allRPAs 10:00:00:00:c9:5e:af:8f 50:01:24:82:00:10:4a:a0 50:01:24:82:00:10:ab:a4 50:01:24:82:01:30:be:9f 50:01:24:82:01:30:e3:9d zone: L_allRPAs_allSTORAGE 50:01:24:82:00:10:4a:a0 50:01:24:82:00:10:ab:a4 50:01:24:82:01:30:be:9f 50:01:24:82:01:30:e3:9d 50:06:01:60:3c:e0:09:c1
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 22 of 223
Step Action 3 TARGET Zones. Create a new zone. As per EMC Best Practices, you will add only two
members to this zone. The first RPA port connected to this fabric and the host HBA. Use one of the RPA WWNs saved on step 1. Command Syntax: zonecreate “zone name”, “RPA3 PWWN 0; Host HBA PWWN” Command Example: zonecreate "rpa3_p0_win2a", "50:01:24:82:01:30:75:a4; 10:00:00:00:c9:69:e6:92" There are many ways to find out the host HBA WWN. Choose the one you are more comfortable with. Hint, hosts already have zones created to the storage ports, looking at an existing zone may be the easiest way to discover that. Run “cfgactvshow” to list the active zones in the configuration. But you can cross check that by looking direct at the host operating system.
Now create additional zones using the same host HBA with all the other RPA ports in this fabric. At the end, there should be four similar zones (you have created the first), with two members each, being one RPA port and the host HBA. Please repeat step 3. Give different names to the new zones. Use the RPA WWNs saved on step 1.
4 INITIATOR Zones. Create another new zone. As per EMC Best Practices, you will add only two members to this zone. The first one is an RPA port connected to this fabric and the second one the first storage port. Command Syntax: zonecreate “zone name”, “RPA3 PWWN 0; Storage Port PWWN a” Command Example: zonecreate "rpa3_p0_spA2", "50:01:24:82:01:30:75:a4; 50:06:01:62:3c:e0:05:91" Verify which storage port to use in the lab environment reference document. There are many ways to find out the storage port WWN. Choose the one you are more comfortable with. Hint, storage ports already have zones created on this switch (zoned with hosts), looking at an existing zone may be the easiest way to discover that. Run “cfgactvshow” to list the active zones in the configuration.
4a Create another new zone. You will add the same RPA port used in the previous step and the second storage port connected to this fabric. Command Syntax: zonecreate “zone name”, “RPA3 PWWN 0; Storage Port PWWN b”
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 23 of 223
Step Action Command Example: zonecreate "rpa3_p0_spB3", "50:01:24:82:01:30:75:a4; 50:06:01:6b:3c:e0:05:91" Note: This second zone is required when Active/Passive storage arrays are present. In this case, this second zone will allow for CLARiiON trespassing of LUNs. In a Symmetrix environment you might not need this zone if only one storage port is available.
Now create additional zones using the same storage ports with all the other RPA ports in this fabric. At the end, there should be eight (four if DMX in place) similar zones (you have created the first), with two members each, being one RPA port and one of the storage ports. Please repeat steps 4 and 4a. Give different names to the new zones. Use the RPA WWNs saved on step 1.
5 Add the new zones to the active configuration and enable the new configuration. Command Syntax: cfgadd “configuration name”, “zone name; zone name; ...” cfgenable “configuration name” Command Syntax: cfgadd “s1_7600a_cfg”, “rpa3_p0_win2a; rpa3_p2_win2a; rpa4_p0_win2a; rpa4_p2_win2a; rpa3_p0_spA2; rpa3_p0_spB3; rpa3_p2_spA2; rpa3_p2_spB3; rpa4_p0_spA2; rpa4_p0_spB3; rpa4_p2_spA2; rpa4_p2_spB3” cfgenable “s1_7600a_cfg”
6 Display the active configuration. You should see your new zones in the active configuration. Command Example: Cfgactvshow
Note that this lab exercise only showed the steps to create zones from the RPAs to a host and zones from the RPAs to the storage ports. You might need to create additional zones for the other hosts and storage ports (if present). Refer to the lab environment reference document.
You have now taken care of zoning on the first fabric at local site. Telnet into your second switch (S1-brocade2_b, S1-brocade4_b etc) at local site and do the same thing (steps 1 to 6) – obviously the zone names and the configuration name may be different there.
7 Check the zoning configuration from the RPAs. Log into each RPA as the boxmgmt user.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 24 of 223
Step Action From the Main menu select Diagnostics > Fibre Channel diagnostics > Detect Fibre Channel targets. Each ports should display the WWNs of the host HBA ports and storage ports.
Important: When you are done with zoning at local site (both switches), move on to remote
site. Telnet into your first and second switches at remote site and do the same thing (steps 1 to 7) – obviously the zone names and the configuration name may be different there.
End of Lab Exercise
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 25 of 223
Part 4: LUN Masking for the RPA – Symmetrix Step Action 1 Before we can replicate volumes using RecoverPoint, we must first present the repository,
journals, and host volumes (replica) to all RPAs at each site. This means a lot of masking. Refer to the lab environment reference guide for repository, journals, and host volumes (replica) information assigned to you RPA cluster group. You will also need information about the storage ports. Finally, WWNs of the RPA ports will be used in this exercise. You can use the syminq or inq command on your hosts to identify the host data (replica) volumes that must be seen by the RPA ports. You will need the Symmetrix device numbers to verify that these volumes are seen by the RPAs.
Command Example: syminq or inq
Sample Output: inq
2 Verify which Symmetrix ports are connected to the host. You will need the director number and port number for the masking commands.
Command Example: symcfg list –connections
Sample Output: symcfg list –connections
# inq ------------------------------------------------------------------------------------------ DEVICE :VEND :PROD :REV :SER NUM :CAP(kb) ------------------------------------------------------------------------------------------ /dev/rdsk/c1t5006048AD5F0F8E1d0s2 :EMC :SYMMETRIX :5772 :5500020341 : 92160 /dev/rdsk/c1t5006048AD5F0F8E1d81s2 :EMC :SYMMETRIX :5772 :5500071341 : 4128000 /dev/rdsk/c1t5006048AD5F0F8E1d82s2 :EMC :SYMMETRIX :5772 :5500072341 : 4128000 /dev/rdsk/c1t5006048AD5F0F8E1d83s2 :EMC :SYMMETRIX :5772 :5500073341 : 4128000 /dev/rdsk/c1t5006048AD5F0F8E1d84s2 :EMC :SYMMETRIX :5772 :5500074341 : 4128000 /dev/rdsk/c1t50012482001063A0d0s2 :KASHYA :KBOX2_S1-BOS1 :0001 : : 983500000 /dev/rdsk/c1t500124820130AAA4d0s2 :KASHYA :KBOX1_S1-BOS1 :0001 : : 983500000 /dev/rdsk/c2t5006048AD5F0F8CFd0s2 :EMC :SYMMETRIX :5772 :5500020480 : 92160 /dev/rdsk/c2t5006048AD5F0F8CFd81s2 :EMC :SYMMETRIX :5772 :5500071480 : 4128000 /dev/rdsk/c2t5006048AD5F0F8CFd82s2 :EMC :SYMMETRIX :5772 :5500072480 : 4128000 /dev/rdsk/c2t5006048AD5F0F8CFd83s2 :EMC :SYMMETRIX :5772 :5500073480 : 4128000 /dev/rdsk/c2t5006048AD5F0F8CFd84s2 :EMC :SYMMETRIX :5772 :5500074480 : 4128000 /dev/rdsk/c2t50012482013063A0d0s2 :KASHYA :KBOX2_S1-BOS1 :0001 : : 983500000 /dev/rdsk/c2t500124820010AAA4d0s2 :KASHYA :KBOX1_S1-BOS1 :0001 : : 983500000 /dev/rdsk/c2t50012482001025A5d0s2 :KASHYA :KBOX2_S1-BOS1 :0001 : : 983500000 /dev/rdsk/c2t500124820130A7A4d0s2 :KASHYA :KBOX1 S1-BOS1 :0001 : : 983500000
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 26 of 223
Step Action
3 Log into the RPA as boxmgmt to get the RPA PWWNs. From the boxmgmt main menu select Diagnostics > Fibre Channel diagnostics > View Fibre Channel details.
View Fibre Channel Details – Sample Output
Port 0 ------------------------------------- |----------------|--------------------| | | | |----------------|--------------------| | Link status | Up | | wwn | 0x50012482001063a0 | | node wwn | 0x50012482001063a1 | | port id | 4f0000 | | operating mode | point to point | | speed | 2 GB | |----------------|--------------------|
Port 0 ------------------------------------- |----------------|--------------------| | | | |----------------|--------------------| | Link status | Up | | wwn | 0x50012482001063a0 | | node wwn | 0x50012482001063a1 | | port id | 4f0000 | | operating mode | point to point | | speed | 2 GB | |----------------|--------------------|
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 27 of 223
Step Action Verify which Symmetrix ports are connected to the RPAs.
Command Example: Symmask list logins –wwn 50012482001063a0
4 Use symcli to mask the replication volumes, journal volumes, and repository volume to the RPA ports. Command Syntax: symmask –sid <Symm S/N> –dir <FA#> –p <Port> –wwn <RPA PWWN> add dev <“DEV#:DEV#”> Command Example: symmask –sid 525 –dir 1c –p 0 –wwn 5001248201b48316 add dev “79:7c”
Sample Output: symmask add dev
Note1: Using a device range (for example: “79:7c”) works for consecutive device numbers. You will need to re-run the command for each device that is not within a range. Note2: You will need to re-run the command for all RPA ports in the fabric zoned to the storage port.
5 After masking the replica volumes, the journal volumes, and the repository volume to the RPA ports, use symcli to refresh the vcmdb. Your masking changes will not take effect until you refresh the database.
Command Example: symmask –sid 525 refresh
6 Verify your changes by listing the masking database.
Command Syntax: symmaskdb list database –dir <FA#> -port <Port#> Command Example: symmaskdb list database –dir 1c –port 0
Sample Output: symmaskdb list database
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 28 of 223
Step Action
7 Log into the RPA as boxmgmt and verify the masking configuration. From the Main menu select Diagnostics > Fibre Channel Diagnostics > Detect Fibre Channel LUNs. You should see the volumes you just masked listed there. Important: Repeat the procedure for all hosts and RPAs at both local and remote sites..
Important: Be sure to perform the following:
• Masking procedure for all RPA ports. o All RPAs shall see all host data volumes (replica) involved in replication as
well as the repository and journal volumes. • Repeat the procedure using RPA ports and storage ports in the second fabric. • Repeat the procedure for site 2 (both fabrics).
End of Lab Exercise
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 29 of 223
Part 5: LUN Masking for the RPA – CLARiiON Step Action 1 Before we can replicate volumes using RecoverPoint, we must first present the repository,
journals, and host volumes (replica) to all RPAs at each site. This means a lot of masking. Refer to the lab environment reference guide for repository, journals, and host volumes (replica) information assigned to you RPA cluster group. You can use the inq or navicli command on your hosts to identify the host volumes (replica) that must be in a CLARiiON storage group with RPA ports. You will need the CLARiiON LUNs information to verify that these volumes are seen by the RPAs.
Command Example: inq or navicli
Sample Output: inq –clar_wwn
2 Launch Navisphere Manager. Refer to the lab environment reference document for the CLARiiON SP IP addresses to be used by your group.
Inquiry utility, Version V7.3-891 (Rev 2.0) (SIL Version V6.5.2.0 (Edit Lev el 891) Copyright (C) by EMC Corporation, all rights reserved. For help type inq -h. ............ ------------------------------------------------------------------------------ DEVICE :VEND :PROD :REV :SER NUM :CAP(kb) ------------------------------------------------------------------------------ \\.\PHYSICALDRIVE0 :( :ST310210A :3.17 : : ----- \\.\PHYSICALDRIVE1 :KASHYA :KBOX1_SOUTHBORO :0001 : : ------ \\.\PHYSICALDRIVE2 :KASHYA :KBOX2_SOUTHBORO :0001 : : ------ \\.\PHYSICALDRIVE3 :KASHYA :KBOX2_SOUTHBORO :0001 : : ------ \\.\PHYSICALDRIVE4 :KASHYA :KBOX1_SOUTHBORO :0001 : : ------ \\.\PHYSICALDRIVE5 :KASHYA :KBOX2_SOUTHBORO :0001 : : ------ \\.\PHYSICALDRIVE6 :KASHYA :KBOX1_SOUTHBORO :0001 : : ------ \\.\PHYSICALDRIVE7 :KASHYA :KBOX2_SOUTHBORO :0001 : : ------ \\.\PHYSICALDRIVE8 :KASHYA :KBOX1_SOUTHBORO :0001 : : ------ \\.\PHYSICALDRIVE9 :DGC :RAID 5 :0428 :0B00005A : 5242880 \\.\PHYSICALDRIVE10 :DGC :RAID 10 :0428 :1F00005C : 2097152 \\.\PHYSICALDRIVE11 :DGC :RAID 10 :0428 :1E00005C : 2097152
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 30 of 223
Step Action 3 Register the RPA ports on the Connectivity Status Window. In the Enterprise Storage
dialog box, navigate to the icon for the storage system whose RPA ports you want to register. Right-click it and select Connectivity Status to open the Connectivity Status dialog box.
4 In Initiator Records, select the HBA that you want to register, and then click Register to open the Register Initiator Record dialog box. The selected HBA must have a No value in the Registered column to enable the Register button. There should be 16 of these entries (which are ‘logged in’). These represent 4 HBAs from each RPA going to 2 CLARiiON SP ports. These entries were created after you finished your zones. Connectivity Status Window
Note: If RPA ports are not listed check your zones on both fabrics and fix any issue they might have. Note: If RPA ports are not ‘logged in’, this is a known and possible issue with CLARiiON arrays. If no I/O happens against the SP ports the initiator will not log in them. To fix this:
• SSH into each RPA as the boxmgmt user. • From the Main menu select Diagnostics > Fibre Channel diagnostics > Detect
Fibre Channel LUNs. • Running this should make the RPAs to log into the CLARiiON SP ports. Therefore,
the initiators ‘logged in’ status should change to ‘Yes’.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 31 of 223
Step Action 5 Verify that the HBA WWN and the SP Port ID are correct and set Initiator Type to
RecoverPoint Appliance. Set Failover Mode to 4. Unless the storage system is connected to a Solaris server, set Unit Serial Number to Array. Enter the Vendor and Model for the initiator. Under “This HBA Belongs to”, click New Host and enter RPA name and address or click Existing Host, and then from the list, select the RPA. Save your changes and close the dialog box to return to the Connectivity Status dialog box.
Register Initiator Window
Note: Alternatively, you can use group edit (previous window) and register all RPAs HBAs as one single entity.
• Initiator type is RecoverPoint Appliance (31). • Failover mode should be set to 4. • Use the management IP of the first RPA to identify the host.
6 You will need to make the host volumes (replica) visible to the RPA ports by adding the volumes to a RPA storage group. View the existing host storage group to verify which volumes are visible. Alternately, you can use inq, syminq or navicli from the host itself.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 32 of 223
Step Action 7 Create a new storage group for the RPAs. You will be adding the host volumes (replica
LUNs), the journal volumes, and the repository volume to this storage group. From the Storage tab of the Enterprise Storage dialog box, navigate to the icon for the storage system on which you want to create the storage group, right-click, and then select Create Storage Group.
Creating a Storage Group
8 Assign a name to your storage group, and click OK or Apply to save your changes.
9 Add the RPAs to the storage group. From the Storage tab of the Enterprise Storage dialog box, navigate to the icon for the storage group to which you want to connect a server, right-click, and then select Connect Hosts.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 33 of 223
Step Action 10 Add the RPAs as Hosts to be Connected. Click Apply to save the changes.
Adding RPAs to the Storage Group
Note: In case you used ‘group edit’ in the RPA HBA registration window (connectivity status), there will be only one RPA listed here.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 34 of 223
Step Action 11 Click on the LUNs tab. Under the Show LUNs: drop down select All. Select all the host
volumes (replica LUNs) to be used for replication, select all the LUNs to be used as journal volumes, and select the LUN to be used as the repository volume. LUN information for each group are listed in the lab environment reference document. Click OK or Apply to save the changes.
Adding LUNs to the Storage Group
12 Expand the Hosts and LUNs icons under your storage group to view the resulting configuration.
13 Log into the RPA as boxmgmt and verify the masking configuration. From the Main menu select Diagnostics > Fibre Channel Diagnostics > Detect Fibre Channel LUNs. The LUNs you just masked to the RPA should be listed there.
Sample Output: Detect Fibre Channel LUNs
|-----|--------|-----|-----------|--------------|--------------|------------| | |Size |Vndr |Product |Name |UID |Site | |-----|--------|-----|-----------|--------------|--------------|------------| | 1. | 3.94GB | EMC | SYMMETRIX | DEV ID: 00A2 | 60,06,04,80, | | | | | | | | 00,01,90,30, | | | | | | | | 11,55,53,30, | | | | | | | | 30,30,41,32 | | | 2. | 3.94GB | EMC | SYMMETRIX | DEV ID: 0073 | 60,06,04,80, | | | | | | | | 00,01,90,30, | | | | | | | | 11,55,53,30, | | | | | | | | 30,30,37,33 | |
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 35 of 223
Step Action Important: Be sure to perform the following:
• Masking procedure for all RPA ports.
o All RPAs shall see all host data volumes (replica) involved in replication as well as the repository and journal volumes.
• Repeat the procedure for site 2.
End of Lab Exercise
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 36 of 223
Lab Exercise 2: Installation of RecoverPoint Appliances
Purpose:
Install the RecoverPoint Appliances using the Installer. Once connected to the RPA over the IP network you will have to configure all RPA parameters for your specified cluster. Once the RPAs are installed they are ready to be configured for replication activities.
Tasks: In this lab, you perform the following tasks: • Connect to the RPAs using the Hardware Installer and perform
configuration of site layout, replication method, site information, IP parameters, and repository volume.
References: • RecoverPoint Student Guide
• RecoverPoint Administrator’s Guide
• RecoverPoint Installation Guide
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 37 of 223
Part 1: RPA Cluster Installation Step Action
1 Log into all RPAs in both site 1 and site 2 as boxmgmt and make sure they allow communications. As previously noted, all RPAs already have management IP so activities can be performed from the classroom. Therefore, you can skip this step.
2 From a Remote Desktop Connection into one of the Windows hosts assigned to your group, launch ‘RecoverPointInstaller.exe’. You should be able to find it in drive C: folder \software\. If you cannot find it, download it from RecoverPoint Software Downloads page on Powerlink.
3 Create a new installation configuration file. Name the file to something convenient and related to the RPA cluster you are installing.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 38 of 223
Step Action 4 Define the system layout.
Number of Sites: 2 Site1 Name: Local Site2 Name: Remote Site to Install Now: Site 1 and Site 2 Replication Protocol: IP Number of RPAs Per Site: 2
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 39 of 223
Step Action 5 Enter Local and Remote Sites IP information.
Please note that the “Site Management IP” and “RPA Management IP” are not the same. Here you should enter the “Site Management IP” (floating), which is an unique address per site.
Refer to lab information sheet.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 40 of 223
Step Action 6 Enter Local and Remote sites RPAs IP information.
Now you enter the RPAs management IPs. Each RPA has two IP addresses: one for management and one for replication.
Refer to lab information sheet.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 41 of 223
Step Action 7 Enter Local and Remote sites general networking details.
Refer to lab information sheet. Notes:
1. Set time zones for both sites. 2. In this lab deployment use the same Primary and Secondary DNS Servers at both sites. 3. Leave the fields for Local Domain blank. 4. NTP Server is set at either Local or Remote site. You cannot configure it at both sites.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 42 of 223
Step Action 8 Leave the Advanced mode window with the default values.
Note for Group4 only: Choose ‘Yes’ in the first option if you have zoned and masked the ESX Server available in your setup. That is OK if you plan on doing those steps later.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 43 of 223
Step Action 9 Login as user boxmgmt. All RPAs in both sites must have management IP address
information set in order to succeed in this step.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 44 of 223
Step Action 10 Move on to next step as zoning and masking tasks have been performed previously.
Note1: When zonings and masking have not been completed yet at this point, if you select the button ‘Zoning and LUN Masking Instructions’ the Installer helps you understand which tasks are required. Note2: The Installer for version 3.1 sometimes do not show all RPA HBA ports information. If you want to validate that there is not issue with RPA HBA ports though, connect into the RPAs using SSH and login as ‘boxmgmt’ user. View Fibre Channel Details from the Diagnostics menu.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 45 of 223
Step Action 11 SAN diagnostics check. Issues with zoning and masking will be verified during this step.
Note: RPAs shall see host and storage ports. Also, RPAs must have access to all volume types: repository, journals and host data volumes.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 46 of 223
Step Action 12 Select the ‘unique’ Repository volume at Local site. The volume should have been
presented to all RPAs at the site via masking. Refer to the lab information sheet to identify the Repository volume assigned to your group.
In our lab environment there is a mix of storage array types. Our replication structure is from CLARiiON to Symmetrix or vice-versa. One important note to make is that the Repository volume size recommendation is 4GB and extra space is not used. Ignore warning messages in case the repository volume assigned to your group is greater than 4GB. Of course, you do not want to do the same at a customer environment as you would be “wasting” storage space.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 47 of 223
Step Action 13 Verify that you have selected the correct volume to be the Repository at Local site. Be
careful, as result of this task the volume will be formatted.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 48 of 223
Step Action 14 Local site RPAs reboot. After repository volume is formatted, all RPAs at the site
reboot. The Installer waits and holds the session until all RPAs are back online with OK status.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 49 of 223
Step Action 15 At this point all RPAs at Local site are installed and configured, now move on to
install and configure RPAs at Remote site. As part of the process of installing and configuring RPAs at Remote site, select and format the unique Repository volume at that site.
16 Remote site RPAs reboot. After repository volume is formatted, all RPAs at the site reboot. The Installer waits and holds the session until all RPAs are back online with OK status.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 50 of 223
Step Action 17 When RPAs at both Local and Remote sites are completed installed and
configured, the RPA cluster is deployed. You can now connect into the RecoverPoint Management Application by either using the given link in the Final Summary window or firing up Internet Explorer and pointing it to the site management IP (floating IP) at either Local or Remote site.
Note: An .xml configuration file is created and saved locally in your service station. Refer to step 3 for the name of the file. This file can be used to reconstruct the RPA cluster, initial install only – repository volumes are formatted if Installer used again. Therefore, it should not be used for RPA cluster changes.
End of Lab Exercise
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 51 of 223
Part 2: RPA Installation – Add RecoverPoint License Step Action
1 Open Internet Explorer and point the browser to the Site Management IP (floating) of either Site 1 or Site 2. Log in as user admin on the RecoverPoint Management Application. Click Yes on the License Problem pop-up - OR Click on the link on the lower left-hand corner of the main window and enter the account ID and license key. There is no activation code for our environment.
First Time Login as admin: Prompt for license information
2 Enter the account ID, company name and contact info then click the Update button to add the license key. You will not be adding software serial ID neither an activation code for the lab environment. Ask your instructor for these information.
Enter the Base License
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 52 of 223
Step Action 3 Enter the license key and click OK.
Update License Key
End of Lab Exercise
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 53 of 223
Lab Exercise 3: Installing Host Splitter Drivers
Purpose:
Perform the installation of the Host Splitter Driver (kdriver) on Windows, Solaris and AIX hosts
Tasks: In this lab, you perform the following tasks: • Installation of the Windows splitter driver • Installation of the Solaris splitter driver • Installation of the AIX splitter driver
References: • Lecture material
• Administrators Guide
• Installation Guide
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 54 of 223
Part 1: Host Splitter Driver Installation - Windows Step Action
1 Log on to your assigned server as “Administrator”.
2 Copy the kdriver installation package to a temporary directory on your server. Be sure to copy a version of the kdriver that supports the RecoverPoint revision on the RPAs.
Note: The lab servers may already have the installation package in C:\software. If not, download it from the RecoverPoint software download page on Powerlink.
3 Extract the contents of the kdriver installation package. To install the kdriver, double-click on the file setup.exe.
kdriver installation package and contents
4 Follow the prompts to install the RecoverPoint Windows-based Splitter Driver.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 55 of 223
Step Action 5 Select Typical when prompted for the Installation Mode to accept the defaults for log file
location and size. To change the log file location and size, select Custom.
6 When zoning and volume presentation has been correctly configured, you should not
receive any errors; your view will be something similar to this:
SAN Diagnostics – No Errors
Note: During the install, the setup.exe program will run diagnostics on the system and the state of its configuration. If the host has not been zoned and does not see SAN connectivity or appliance volumes, these configurations should either be made or rectified.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 56 of 223
Step Action 7 Once the installation completes, click Finish to exit the setup and restart the computer.
Restart Host
8 After the system reboots verify the kdriver is running by checking Services for the KDriverWatchdog service.
KDriverWatchdog Service
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 57 of 223
Step Action 9 Run an inq or syminq command and verify that the host sees the appliance as a disk.
Important: Repeat Steps 1 through 9 for each Windows host in the replication environment (both sites).
End of Lab Exercise
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 58 of 223
Part 2: Host Splitter Driver Installation – Solaris Step Action
1 Log on to your assigned server as “root”.
2 Copy the kdriver installation package to a temporary directory on your server. Be sure to copy a version of the kdriver that supports the RecoverPoint revision on the RPAs. Note: The lab servers may already have the installation package in /software/. If not, download it from the RecoverPoint software download page on Powerlink.
3 Decompress the file.
Command Example: gunzip filename
Decompress File
4 Use pkgadd –d to install the kdriver package:
Beginning Pkgadd Installation
sl—sunl /software/solaris/kdriver pkgadd -d kdriver_3.2.e.131_sol.pkg The following packages are available: 1 kdriver kdriver (sparc) 3.2(k.131) Select package(s) you wish to process (or 'all' to process all packages). (default: all) [?,??,q]:
sl—sunl /software/solaris/kdriver is —la total 19956 drwxr—xr—x 2 root root 512 Sep 8 08:27 drwxr—xr—x 7 root other 512 Sep 8 08:27 —ru—r——r—— 1 root root 10199278 Sep 8 08:28 kdriver_3.2.e.131_sol.pkg.gz sl—sunl /software/solaris/kdriver gunzip kdriver_3.2.e.131_sol.pkg.gz sl—sunl /software/soiaris/kdriver ls —la total 83604 drwxr—xr—x 2 root root 512 Sep 8 08:29 drwxr—xr—x 7 root other 512 Sep 8 08:27 —rw—r——r—— 1 root root 42767360 Sep 8 08:28 kdriver_3.2.e.131_sol.pkg s1—sun1 /software/solaris/kdriver
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 59 of 223
Step Action 5 When prompted, enter y to continue the installation.
6 Correct any errors detected by SAN diagnostics. Note: The system conducts a series of automatic tests that are designed to identify a variety of the most common problems regarding the host configuration in the SAN environment on which RecoverPoint is being installed. It is desirable to identify these problems before installing the kdriver on the host. Upon completion of these tests, a report confirms successful completion of the SAN diagnostics (i.e., no errors or warnings), or details the errors or potential errors (i.e., warnings) detected in the host configuration on the SAN.
7 When prompted, designate the directory in which to store log files and designate the maximum log file size (default is 500 MB). The default destination folder is /kdriver/log.
Setting Log Directory and Size
8 When the installation completes, reboot the host.
Command Example: reboot -- -r
Note: From this point, the kdriver is started automatically any time the system is rebooted. To manually start or stop the kdriver use the /etc/init.d/kdrv start and /etc/init.d/kdrv stop commands. The /etc/init.d/kdrv status command returns the current status of the driver (UP or DOWN).
Do you want to proceed with the installation (y/q)? y Please enter the log files directory (q-quit)? [/kdriver/log] Please enter the log files maximum size [MB] (q-quit)? [500] kconfigure – success installation of <kdriver> was successful.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 60 of 223
Step Action 9 There are several commands that can be run on Solaris to verify that the splitter driver is
installed and running properly. The pkginfo command displays information about software packages that have been installed on the system. Look at the STATUS field to verify the driver was installed.
Command Example: pkginfo –l kdriver
Sample Output: pkginfo –l kdriver
10 The modinfo command will display information about loaded kernel modules.
Command Example: modinfo | grep -i splitter
11 The ps command displays the status of processes running on the system.
Command Example: ps –ef | grep kdriver
Sample Output: ps –ef | grep kdriver
s1-sun1 / ps –ef | grep kdriver root 462 461 0 10:49:50 ? 0:03 /kdriver/bin/kdriver root 461 1 0 10:49:50 ? 0:00 /kdriver/bin/kdrive_swd 60 /kdriver 1 60 0
root 463 461 0 10:49:50 ? 0:01 /kdriver/bin/hlr host
s1-sun1 / modinfo | grep –I splitter 286 78666000 2a450 282 1 splitter (KashyaSplitter)
sl—sunl /software/solaris/kdriver pkginfo –l kdriver PKGINST: kdriver NAME: kdriver CATEGORY: application ARCH: sparc VERSION: 3.2 BASEDIR: / PSTAMP: sol15920080219184442 INSTDATE: Sep 2 2009 10:43 STATUS: completely installed FILES: 118 installed pathnames 13 directories 48 executables 83192 blocks used (approx)
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 61 of 223
Step Action 12 Run an inq or syminq command and verify that the host sees the appliance as a disk.
Sample Output: inq
Important: Repeat these steps for each Solaris host in the replication environment (both sites)
End of Lab Exercise
s1-sun1 / syminq | grep KASHYA /dev/rdsk/c2t0d0s2 KASHYA KBOX2_S1BOS1 0001 N/A 983500000 /dev/rdsk/c2t1d0s2 KASHYA KBOX1_S1BOS1 0001 N/A 983500000 /dev/rdsk/c4t0d0s2 KASHYA KBOX1_S1BOS1 0001 N/A 983500000 /dev/rdsk/c4t1d0s2 KASHYA KBOX2_S1BOS1 0001 N/A 983500000 /dev/vx/rdmp/c2t0d0s2 KASHYA KBOX2_S1BOS1 0001 N/A 983500000 /dev/vx/rdmp/c2t1d0s2 KASHYA KBOX1_S1BOS1 0001 N/A 983500000 /dev/vx/rdmp/c4t0d0s2 KASHYA KBOX1_S1BOS1 0001 N/A 983500000 /dev/vx/rdmp/c4t1d0s2 KASHYA KBOX1 S1BOS1 0001 N/A 983500000
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 62 of 223
Part 3: Host Splitter Driver Installation – AIX Step Action
1 Copy the installation file to the AIX host. Use the installp command to install the kdriver on both the Site 1 and Site 2 hosts. Be sure to copy a version of the kdriver that supports the RecoverPoint revision on the RPAs. Note: The lab servers may already have the installation package in /software/. If not, download it from the RecoverPoint software download page on Powerlink.
# ls .toc EMC_RecoverPoint_driver_aix52_53_rel3.2.SP1_g.63_md5_b07fc87f45440042f13574ee4a54f9ba.bff # installp -agXd . all +-----------------------------------------------------------------------------+ Pre-installation Verification... +-----------------------------------------------------------------------------+ Verifying selections...done Verifying requisites...done Results... SUCCESSES --------- Filesets listed in this section passed pre-installation verification and will be installed. Selected Filesets ----------------- kdriver.bin 2.0.0.0 # kdriver kdriver.hlr 2.0.0.0 # kdriver kdriver.info_collector 2.0.0.0 # kdriver kdriver.install 2.0.0.0 # kdriver kdriver.kutils 2.0.0.0 # kdriver kdriver.lib 2.0.0.0 # kdriver kdriver.log 2.0.0.0 # kdriver kdriver.modules 2.0.0.0 # kdriver kdriver.tweak 2.0.0.0 # kdriver
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 63 of 223
Step Action 2 Examine the install output to verify the splitter was successfully defined.
3 After the installation completes, run the rc.kdrv configure command to configure the splitter.
# rc.kdrv configure _ __ _____ _ | |/ /| __ \ (_) | ' / | | | | _ __ _ __ __ ___ _ __ | < | | | || '__|| |\ \ / // _ \| '__| | . \ | |__| || | | | \ V /| __/| | |_|\_\|_____/ |_| |_| \_/ \___||_| Running kconfigure Installing RecoverPoint.AIX.5.2_5.3.1.0.0.0 from /kdriver/install/ RecoverPoint.AIX.5.2_5.3.1.0.0.0 alredy installed Searching for RecoverPoint devices... Loading the 64 bit module... Running SAN diagnostics. This may take a few moments...
---------- COPYRIGHT NOTICE ---------- This software is the property of EMC Corp. Copyright (C) 2004 - 2007 EMC Corp. All rights reserved. . . . . . << End of copyright notice for kdriver >>. . . . Creating the splitter device splitter0 Defined Adding the Config_Rules to the ODM Please run rc.kdrv configure Creating link for /usr/sbin Finished processing all filesets. (Total time: 16 secs).
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 64 of 223
Step Action 4 Examine the installation summary of installed components.
5 When prompted, enter Y to continue and select the location for the kdriver log and log file size. SAN diagnostic information: 1 errors: 1. Number of devices without UID (2) exceeds maximum (1) 3 warnings: 1. Found device with no UID: dev=/dev/rhdisk0 channel=0 target=n/a LUN=n/a vendor=IBM H0 product=HUS103073FL3800. Contact technical support if you wish to replicate this LUN. 2. Found device with no UID: dev=/dev/rhdisk1 channel=0 target=n/a LUN=n/a vendor=IBM H0 product=HUS103073FL3800. Contact technical support if you wish to replicate this LUN. 3. Host cannot see RPA1 Total=4 Do you want to proceed with the installation (y/q)? y Please enter the log files directory: (q-quit) [/kdriver/log] Please enter the log files maximum size (in MB): (q-quit) [500] 10 validating space on logs FS.... free space: 3825.33203125MB OK /usr/bin/gzip Formatting the volume database: [done] Formatting the kutils database: [done] Formatting the splitter config file: [done] Initializing the splitter: [done] Backup old inittab file: [done] Backup old inittab file: [done] Register KDriver to load at boot time: [done] Register KDriver shutdown notifier: Added Kashya changes to file '/etc/rc shutdown' successfully
Installation Summary -------------------- Name Level Part Event Result ------------------------------------------------------------------------------- kdriver.tweak 2.0.0.0 USR APPLY SUCCESS kdriver.modules 2.0.0.0 USR APPLY SUCCESS kdriver.log 2.0.0.0 USR APPLY SUCCESS kdriver.lib 2.0.0.0 USR APPLY SUCCESS kdriver.kutils 2.0.0.0 USR APPLY SUCCESS kdriver.install 2.0.0.0 USR APPLY SUCCESS kdriver.info_collector 2.0.0.0 USR APPLY SUCCESS kdriver.hlr 2.0.0.0 USR APPLY SUCCESS kdriver.bin 2.0.0.0 USR APPLY SUCCESS kdriver.tweak 2.0.0.0 ROOT APPLY SUCCESS kdriver.modules 2.0.0.0 ROOT APPLY SUCCESS kdriver.log 2.0.0.0 ROOT APPLY SUCCESS kdriver.lib 2.0.0.0 ROOT APPLY SUCCESS kdriver.kutils 2.0.0.0 ROOT APPLY SUCCESS kdriver.install 2.0.0.0 ROOT APPLY SUCCESS
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 65 of 223
Step Action 6 Run the rc.kdrv start command to start the kdriver.
End of Lab Exercise
# rc.kdrv start Starting the KDriver daemon.... # ps -ef | grep kdriver root 213232 1 0 10:49:50 - 0:00 /kdriver/bin/kdriver_swd 0 /kdriver 1 60 1 root 372906 213232 0 10:49:50 - 0:01 /kdriver/bin/hlr_host root 393346 213232 0 10:49:50 - 0:01 /kdriver/bin/kdriver
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 66 of 223
Part 4: Add Splitters Using the RecoverPoint Management Application
Step Action
1 Open Internet Explorer and point the browser to Site Management IP (floating) of either site 1 or site 2. Login as user admin. Using the RecoverPoint Management Application, highlight Splitters in the tree panel. From the Splitter menu, add the new splitters for Site 1 and Site 2.
2 From the Add New Splitter window select the host splitter driver to be added. Click Next.
Note1: If your host splitter driver (kdriver) is not showing up, the most likely cause is zoning issue. Check the zones from RPA ports to your host HBAs. Make sure that you see the RPAs from your host when you run ‘inq’. Note2: Another possible issue is the host splitter driver (kdriver) not being installed nor running on the host. Check this!
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 67 of 223
Step Action 3 Review the selected host splitter drivers to be added to the application and click Finish.
Remember to repeat this procedure for all hosts involved in replication at both local and remote sites.
4 Verify the status of the newly added splitters.
Splitter Status
End of Lab Exercise
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 68 of 223
Lab Exercise 4: Replication Configuration
Purpose:
Create first basic Consistency Group and perform first replication job.
Tasks: In this lab, you perform the following tasks: • Creating journal and replication volumes • Assigning volumes • Attaching volumes to host splitters • Creating Consistency Groups • Performing first base-line replication
References: • Lecture material • Administrators Guide • Installation Guide
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 69 of 223
Part 1: Create Consistency Group Step Action
1 Launch the RecoverPoint Management Application and login as user admin. Highlight Consistency Groups in the tree panel. Click on the Groups menu and select Add New Group. Or right-click the Consistency Group category in the tree-panel and select Add Group.
2 In the Consistency Group Wizard, name the consistency group and select a preferred RPA. Click
Next. Consistency Group Wizard
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 70 of 223
Step Action 3 Name the production copy and leave the Advanced settings at the default settings. Click Next
Production Copy Settings
4 Enter a name for the remote copy and leave the Advanced settings at the default settings. Click Next.
Replica Copies Settings
Do not create a local copy at this moment. Therefore, do not name it.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 71 of 223
Step Action 5 Under Replication Sets step, select one or more production volumes to replicate. Refer to lab
information sheet for devices to be used. Replication Sets
6.a Select a volume from Production Volume of Replication Set 1 area at the top of the screen,
then, from the Volumes at pod2 that can be Added to Replication Set 1 list, select the volume that you want to replicate the specified production volume to.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 72 of 223
Step Action 6.b Review the information in the Replication Set Configuration screen. Click Next.
Replication Sets Summary
7.a Select the volumes from Volumes at pod1 that can be Added to Production Copy Journal
area that you want to add to the journal at the copy site. Multiple volumes can be selected. Production Journal Volumes
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 73 of 223
Step Action 7b Select the volumes from Volumes at pod2 that can be Added to Replica Copies Journal area
that you want to add to the journal at the remote site. Multiple volumes can be selected. Remote Journal Volumes
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 74 of 223
Step Action 8 Review the settings in the Create Consistency Group screen, and verify that they are correct.
Click Finish to complete the consistency group configuration.
Note: Uncheck the Start data transfer immediately checkbox.
If host splitter driver has already been installed in the previous lab and volumes have been auto-attached to splitters during the consistency group creation wizard, please skip steps 9 through 12 and move on to step 13. Otherwise, go back to previous lab and install host splitter driver on the hosts assigned to your group. When that is done, please continue with this lab step 9.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 75 of 223
Step Action 9 a) Add the splitters to the RecoverPoint Management Application. Select the more convenient
way.
b) Attach the replication volumes to the splitters. Highlight each splitter and click on the
Splitter Properties icon (or just double click on it).
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 76 of 223
Step Action 10 From the Splitter Properties window, click Attach and select the volume to attach. Click OK.
Click OK again to close the Splitter Properties window. Remember to attach the site 1 and site 2 replication volumes to the corresponding splitter.
Splitter Properties
11 Make sure consistency group is enabled, if not enable it and begin data transfer. Highlight the consistency group and click on the Enable Group icon.
Enable Consistency Group
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 77 of 223
Step Action 12 Click Yes to enable the consistency group.
Enable Group: Confirmation Message
13 From the tree-panel, highlight the consistency group to view the status window. The consistency
group will initialize and then display a transfer status of Active. Consistency Group Status
Note: Notice how the remote site storage state shows ‘N/A’. It should display ‘No access’ instead, the issue here is that the host splitter driver (kdriver) has not been attached to the volumes on the remote copy. If you see the same behavior in your devices, please go ahead and fix it. By now you should know how to do it.
End of Lab Exercise
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 78 of 223
Lab Exercise 5: Managing Replication Jobs
Purpose:
Manage RecoverPoint Replication jobs. Access remote site images; redirect processing, pausing replication, resuming replication.
Tasks: In this lab, you perform the following tasks: • Pausing replication • Accessing remote images • Creating Bookmark image • Accessing Bookmark • Resuming replication
References: • Lecture material • Administrator’s Guide • CLI Guide • Best Practices for Windows Replication
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 79 of 223
Part 1: CRR Image Access Step Action
1 Launch the RecoverPoint Management Application and login as user admin, you probably have a session opened already. Disable the consistency group. From the consistency group status window click on the Disable Group icon. Click Yes to confirm.
Disable Consistency Group
2 Mount the replication volumes (in your consistency group) on hosts at both the local and remote site (site 1 and site 2). It might be necessary to make the volumes available to the OS (format volume, create file system, assign drive letter). Please refer to the lab information sheet for volumes information.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 80 of 223
Step Action 3 Umount the replication volumes at the remote site (site 2) using kutils. The kutils utility
was installed automatically when you installed RecoverPoint splitter on your hosts .
It is necessary to be in the correct path (default C:\Program Files\KDriver\kutils) to run kutils commands. A return message (success or failure) is always expected. The kutils utility is case sensitive.
Command Example
To unmount a device: C:\program files\kdriver\kutils> kutils umount E: Unmounting drive E:... unmounted from "\\?\Volume{33b4a391-26af-11d9-b57b-505054503030}\" This command is available only on hosts running Windows. For Unix host use its native umount utility. For Windows 2003 and later, use the mountvol.exe, which comes with the Windows operating system. Command Example To umount a device: C:\>mountvol E: /p
4 Re-enable the consistency group by clicking on the Enable Group icon. Click Yes to confirm.
Enable Consistency Group
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 81 of 223
Step Action 5 On the site 1 host, copy some files to the production volume.
6 Flush file system on the replication volumes at the local site (site 1) using kutils.
It is necessary to be in the correct path (default C:\Program Files\KDriver\kutils) to run kutils commands. A return message (success or failure) is always expected. The kutils utility is case sensitive.
Command Example To initiate an OS-flush of the file system on the device designated as drive E: C:\program files\kdriver\kutils> kutils flushFS E: Flushing buffers for drive E:... Flushed. This command is available only on hosts running Windows.
7 Create a bookmark by click on the Create Bookmark icon. Name the bookmark and click OK.
Create Bookmark
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 82 of 223
Step Action 8 Access the bookmark by clicking on the drop down menu under the remote replica volume
to open the Enable Image Access wizard. Click Select and image from the list. Click Next.
Image Access
9 Select your bookmark and click Next.
Select Image
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 83 of 223
Step Action 10 Select Logged access (physical) and click Next.
Image Access Mode
11 Review the content and click Finish on the Summary window.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 84 of 223
Step Action 12 Once the bookmark is fully distributed and the remote site storage status changes to
“Logged access (0% full)”, login into the remote server (site 2) and mount the replication volume using kutils.
Image Access Status
12a Command Example
To mount a device: C:\program files\kdriver\kutils> kutils mount E: Mounting drive E: as "\\?\Volume{33b4a391-26af-11d9-b57b-505054503030}\"... Mounted. This command is available only on hosts running Windows. For Unix hosts use its native mount utility. For Windows 2003 and later, use the mountvol.exe, which comes with the Windows operating system. Command Example To mount a device: C:\> mountvol E: \\?\Volume{4b3d346c-c4cd-11dd-b127-001aa0a858f4}\ Type ‘mountvol’ alone to get a list of the possible volume names. Copy and paste the volume name into the command line after the driver letter of the volume you are trying to mount.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 85 of 223
Step Action 13 Access the volume on the remote server (site 2) and verify that it has the same data you
copied into the production server volume (site 1). This should be a proof of concept. Replication is working! If not, you probably missed a step. Go back and check: a) Local and Remote servers have host splitter driver installed and running. b) Local and remote volumes have been attached to splitters. c) You have copied data to the production volume (site 1). d) You have flushed file system of production server volume (site 1) before creating the
bookmark. e) You are using the correct volumes! Make sure that the volumes within your consistency
group are the ones you are copying data to, mounting and umounting.
14 Umount the replication volumes at the remote site (site 2) using kutils. It is necessary to be in the correct path (default C:\Program Files\KDriver\kutils) to run kutils commands. A return message (success or failure) is always expected. The kutils utility is case sensitive.
Command Example
To unmount a device: C:\program files\kdriver\kutils> kutils umount E: Unmounting drive E:... unmounted from "\\?\Volume{33b4a391-26af-11d9-b57b-505054503030}\" This command is available only on hosts running Windows. For Unix host use its native umount utility. For Windows 2003 and later, use the mountvol.exe, which comes with the Windows operating system Command Example To umount a device: C:\>mountvol E: /p
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 86 of 223
Step Action 15 Disable the image access by clicking the dropdown menu under the remote copy (replica)
and selecting Disable Image Access.
Disable Image Access
16 Click Yes on the warning message.
Disable Image Access Warning
Notice how data distribution resumes and the remote site storage status goes back to “No access”. The consistency group is in its previous state, replicating from local site to remote site.
End of Lab Exercise
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 87 of 223
Part 2: CRR Failover/Failback Step Action
1 Umount the replication volumes at the local site (site 1) using kutils. It is necessary to be in the correct path (default C:\Program Files\KDriver\kutils)to run kutils commands. A return message (success or failure) is always expected. The kutils utility is case sensitive.
Command Example
To unmount a device: C:\program files\kdriver\kutils> kutils umount E: Unmounting drive E:... unmounted from "\\?\Volume{33b4a391-26af-11d9-b57b-505054503030}\" This command is available only on hosts running Windows. For Unix host use its native umount utility. For Windows 2003 and later, use the mountvol.exe, which comes with the Windows operating system Command Example To umount a device: C:\>mountvol E: /p
2 Create a bookmark image.
Create Bookmark
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 88 of 223
Step Action 3 Access the new bookmark image. The screenshot does not show the entire procedure, so
enable physical mode and finish the wizard accepting defaults.
Image Access
4 Once the bookmark is fully distributed, perform a failover from the dropdown menu below the
remote replica. Failover Operation
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 89 of 223
Step Action 5 Click Yes to confirm the failover.
Failover Confirmation
6 Click Yes on the warning message. Note that when you perform a failover the journal volume
is erased.
Failover Warning
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 90 of 223
Step Action 7 View the status of the consistency group after the failover. The remote replica is moved to the
left side of the status screen and its role is changed to Production Source.
Failover Status
7a Mount the replication volumes at the “new local” (temporally production) site (site 2, which is on the left side now) using either kutils or mountvol. Copy new data on this server’s volume.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 91 of 223
Step Action 8 In the three panel, select the local copy (which is temporally remote) and display the journal.
Notice the journal is now being updated on the local side.
Local Volume Journal
9 Using either kutils or mountvol, umount the replication volumes at the “new local” (temp
production) site (site 2).
10 Create a new bookmark to perform a failback.
Create Bookmark
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 92 of 223
Step Action 11 Access the new bookmark image.
Image Access
12 Once the bookmark is fully distributed, perform a failover from the dropdown menu below the
remote copy (the original local production site). Failover Operation
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 93 of 223
Step Action 13 Click Yes to confirm the failover.
Failover Confirmation
14 Click Yes on the warning message. Note that when you perform a failover the journal volume
is erased.
Failover Warning
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 94 of 223
Step Action 15 View the consistency group status. Note that the original production copy is moved back to
the left side of the screen and its role is now the Production Source. The opposite happens to the remote copy.
Consistency Group Status
Also, notice that there’s no ‘failback’ command. Only the failback concept applies as the command is always ‘failover’. Failover operations flips the role of the copies every time it is invoked.
End of Lab Exercise
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 95 of 223
Part 3: CRR Recovery Production Step Action
1 Umount the replication volumes at the local site (site 1) using kutils. It is necessary to be in the correct path (default C:\Program Files\KDriver\kutils)to run kutils commands. A return message (success or failure) is always expected. The kutils utility is case sensitive.
Command Example
To unmount a device: C:\program files\kdriver\kutils> kutils umount E: Unmounting drive E:... unmounted from "\\?\Volume{33b4a391-26af-11d9-b57b-505054503030}\" This command is available only on hosts running Windows. For Unix host use its native umount utility. For Windows 2003 and later, use the mountvol.exe, which comes with the Windows operating system Command Example To umount a device: C:\>mountvol E: /p
2 Create a new bookmark to use for the restore operation.
Create Bookmark
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 96 of 223
Step Action 3 Access the new bookmark image.
Image Access
4 Using either kutils or mountvol, mount the remote replica volume and copy new files.
5 Using either kutils or mountvol, umount the replication volumes at the remote site (site 2).
6 Once the bookmark is fully distributed, click on the drop down menu under the remote replica and select Recover production.
Recover Product
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 97 of 223
Step Action 7 Click Yes to confirm the recover operation.
Confirm Recover
8 Click Yes on the warning message. Notice the recover operation block host access to the
remote replica.
Recover Warning
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 98 of 223
Step Action 9 View the consistency group status. Notice the production volume is being restored and the
remove journal is being preserved. Neither volume is currently accessible to a host.
Consistency Group Status
10 Wait for the completion of the restore process. Create a new bookmark to view the contents
of the restored production volume.
Create Bookmark
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 99 of 223
Step Action 11 Access the new bookmark image from the dropdown menu under the production volume.
Access Bookmark
12 Mount the local volume (site 1) and view the updated contents.
13 Resume production operations using the updated production volume by clicking the Resume Production icon under the production volume.
Resume Production
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 100 of 223
Step Action 14 Click Yes to confirm the resume operation.
Resume Confirmation
15 Click Yes on the warning message. Note that the local journal will be erased.
Resume Warning
16 Click on the remote volume and display the remote journal. Notice the remote journal is
still preserved. Remote Journal
End of Lab Exercise
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 101 of 223
Lab Exercise 6: Working with Consistency Group Policies
Purpose:
Modify consistency group policies and observe resulting replication statistics.
Tasks: In this lab, you perform the following tasks: • Modify CG Policies • Copy data to production volumes • Monitor replication statistics to observe effect of policy changes
References: • Lecture material
• Administrator’s Guide
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 102 of 223
Part 1: Consistency Group Policies Step Action
1 Highlight your consistency group and click on the Policy tab. Under General Settings change the Preferred RPA for your cg and click Apply.
Change Preferred RPA
2 Click Yes to confirm the warning message. Notice the data transfer is paused while the preferred RPA is changed.
Warning Message
3 Click on the Status tab to monitor the cg status while the preferred RPA changes.
Consistency Group Status
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 103 of 223
Step Action 4 Click on the Policy tab. Under Compression, change the compression level from 10 to 1.
Click Yes on the warning message.
Change Compression
5 Identify a consistent set of data that you can use to copy to the production volume for the remainder of this lab. Start a copy to the production volume. Click on the Statistics tab and monitor the Replication Performance statistics with compression set to 1. Note: For each lab step, make a note of the scale for each chart. This makes it easier to compare results.
Replication Performance Statistics
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 104 of 223
Step Action 6 Click on the Policy tab and set the compression level back to 10.
Change Compression
7 Start another copy of the same data to the production volume. Be sure to copy the data to a new directory on the production volume (or remove the original copy). Monitor the Replication Performance Statistics during the copy.
Replication Performance Statistics
8 Click on the Policy tab. Under Protection change the Minimize setting from Lag to
Bandwidth.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 105 of 223
Step Action 9 Copy your data to a new directory on your production volume and monitor the Replication
Performance Statistics.
Replication Performance Statistics
10 Click on the Policy tab. Under Protection, set the RPO Specifications for Lag to 50 MB.
RPO Specifications
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 106 of 223
Step Action 11 Copy your data to a new directory on your production volume and monitor the Replication
Performance Statistics.
Replication Performance Statistics
12 Click on the Policy tab and set Optimization back to Lag. Under Protection set the RPO
Specifications back to System Optimized. Copy your data to a new directory on your production volume and monitor the Replication Performance Statistics.
Replication Performance Statistics
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 107 of 223
Step Action 13 Click on the Policy tab. Under Advanced – Remote Link set the Snapshot Granularity to
Fixed (per write).
Snapshot Granularity
14 Copy your data to a new directory on your production volume. Highlight the remote volume
and click on the Journal tab. Notice the Sample Images.
Journal Sample Images
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 108 of 223
Step Action 15 Highlight your consistency group and click on the Statistics tab. Monitor the Replication
Performance.
Replication Performance Statistics
16 Click on the Policy tab. Under Advanced - Remote Link, set the Snapshot Granularity to
Fixed (per second).
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 109 of 223
Step Action 17 Copy your data to a new directory on your production volume. Highlight the remote volume
and click on the Journal tab. Notice the Sample Images.
Journal Sample Images
18 Highlight your consistency group and click on the Statistics tab. Monitor the Replication Performance.
Replication Performance Statistics
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 110 of 223
Step Action 19 Highlight the remote volume and click on the Policy tab. Under Journal, set the Maximum
Journal Lag from Unlimited to Limited, 12 MB.
Remote Volume Policy
20 Copy your data to a new directory on your production volume. Highlight the remote volume
and click on the Journal tab. Notice the Journal Lag and Current Image.
Journal Lag
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 111 of 223
Step Action 21 Highlight your consistency group and click on the Statistics tab. Monitor the Replication
Performance.
Replication Performance Statistics
End of Lab Exercise
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 112 of 223
Lab Exercise 7: Install and Configure CLARiiON Array Based Splitter
Purpose:
Install the splitter patch and enabler on a CX4 array in order to integrate it with a RecoverPoint cluster.
Tasks: In this lab, you perform the following tasks: • Verify array compatibility • Installed the latest Splitter Engine • Install the RecoverPoint Splitter Enabler • Zone the CX to the RPA cluster • Connect the RPAs to the array • Present LUNs to be protected to the RPA cluster by including them
into required Storage groups • Add the CX4 as a splitter to an RPA cluster
References: • EMC RecoverPoint Installation Guide
• EMC RecoverPoint Administrator’s Guide
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 113 of 223
Part 1: Software tasks on the CX array Step Action
1 Verify that the RPAs have been installed with the RecoverPoint operating environment. The RPAs should be installed and ready for configuration before attempting to attach the CX as a splitter. Verify that the CX4 to become a splitter has at a minimum of FLARE28. Using Navisphere manager: Right-click on the array and select Properties.
2 Once the properties window is displayed, check the software revision found within the software tab. If the version is less than FLARE28, upgrade the array. E.G.:
The upgrade procedure for FLARE is not shown here.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 114 of 223
Step Action 3 With the array at the appropriate FLARE version, it is time to proceed with installing the
RecoverPoint Splitter Engine and Enabler. At the time of this writing, these are the versions for these components: CX4 - 04.28.006.6.002 (engine) RPSplitterEnabler-01.01.5.002-xpfree.ena The RecoverPoint Splitter is available as a patch on Powerlink and is not cut in to the manufacturing code for the CLARiiON. You need to have this software component on the management station used to configure and upgrade the CLARiiON (e.g. your laptop).
4 Using Navisphere Taskbar, proceed with installing the Splitter engine and enabler. If you do not have the Taskbar, it is available from: www.cs.isus.emc.com - > CLARiiON Information -> Software Downloads This next section consists of several steps- installing the splitter engine and enabler.
5a Connect to the array using the Taskbar, login using the standard administrator login and password. Once logged in, select to perform a custom install:
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 115 of 223
Step Action 5b After clicking Next, browse to select the software to install. This is shown in the lower-red
box within this graphic:
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 116 of 223
Step Action 5c After selecting the software, it is uploaded from the management workstation to the array.
After this process completes, the array performs a series of checks to ensure that the software upgrade procedure occurs without issue.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 117 of 223
Step Action 5d Once the array checks have completed, the software is installed. It is fine to install both the
engine and the enabler at the same time. The software is installed to both SPs and reboots are a normal part of the process. Proceeding with the installation:
6 Select Finish to complete the installation.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 118 of 223
Part 2: Provisioning RPAs Step Action
1 For all nodes within the RPA cluster, the CX must be able to perform both reading and writing to the RPAs. Zone all RPA ports with all CX SP ports. Ports to exclude in zoning are any ports involved in a MirrorView configuration. Zoning steps are not shown here. It is required to create zones containing all RPA ports connected to one single CX port on the array. If there are two CX ports per fabric, then you need to create two separate zones. For example:
2 Connect the RPAs to the CX. This is done by manually registering the RPAs with the CX
array. For each RPA port, register as Initiator Type = RecoverPoint Appliance and Failover mode 4:
Zone: RPAs12_CX_A0 50:01:24:82:00:93:f7:1c 50:01:24:82:01:aa:2a:50 RPA PORTS 50:01:24:82:00:93:7b:21 50:01:24:82:01:aa:dc:4d 50:06:01:60:3c:e0:0d:a1 CX SPA0 PORTS Zone: RPAs12_CX1_B1 50:01:24:82:01:aa:dc:4d 50:01:24:82:00:93:7b:21 RPA PORTS 50:01:24:82:01:aa:2a:50 50:01:24:82:00:93:f7:1c 50:06:01:69:3c:e0:0d:a1 CX SPB1 PORTS
4
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 119 of 223
Step Action 3 Providing LUN masking for the RPAs can be done in a matter of different ways. The
minimum configuration is the following: • The RPAs must have at least one LUN presented to them/the cluster. • The RPAs must be added to storage groups or have LUNs from storage groups
presented to the RPA Storage Group in order for the RPAs to replicate these LUNs. In our example, we show creating a SG for the RPAs and adding LUNs which is then protected with RecoverPoint. First, create a SG for RecoverPoint, provide one LUN and connect the RPAs to this SG. For example:
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 120 of 223
Step Action 4 Assign at least one LUN and LUNs to be protected (from other SGs) to the RPA SG:
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 121 of 223
Part 3: Add CLARiiON Splitter to the RecoverPoint Cluster Step Action
1 Once the above steps are completed, you should be able to add the CLARiiON as a new splitter within a RecoverPoint cluster. From the RPA GUI, right-click on the Splitter entry in the navigation pane and then select Add New Splitter.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 122 of 223
Step Action 2 You are presented with a list of sites and available splitters. Choose the site/sites with the
CLARiiON splitters and select to add them to the RPA environment.
Add CLARiiON credentials if you are asked to. This will allow for logs to be retained, they are automatically retrieved and stored in the RPA cluster.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 123 of 223
Step Action 3 After you select the splitters and click OK, you see them registered with the RPA GUI:
You can now proceed with configuring replication tasks within the RPA GUI. The last and final step is attaching volumes to the splitters. Basically you need to attach each LUN that exists in a replication set with a CLARiiON splitter.
End of Lab Exercise
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 124 of 223
This page intentionally left blank.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 125 of 223
Lab Exercise 8: Migrate to Cisco SANTap Fabric Splitter
Purpose:
Migrate from kdriver to fabric splitting with SANTap. This procedure must be done while the host still sees and has access to storage. Hosts in this environment have dual paths to storage. Additionally, they have multi-pathing software (e.g. PowerPath). With this type of environment, this procedure can theoretically be done online with no interruption to existing Host-to-Storage (H2S) connectivity.
Tasks: In this lab, you perform the following tasks: • Inventory the environment and analyze current state and what needs
to be upgraded. Qualify the environment. • Stop consistency groups currently replicating and detach volumes
from kdriver splitter. • Stop kdriver on the host- either remove or change configuration so
that it does not start upon reboot. • Take necessary backups of the SAN environment. • Take one path down to make changes to the configuration and deploy
SANTap. • Disable IVR (if applicable to your environment) • Make necessary VSAN changes • Deploy SANTap • Bring fabric back online, verify that host still has access to storage • Repeat configuration changes/deployment on all switches in the
environment • Once SAN environment configuration changes are complete restore
consistency groups with fabric splitter
References: • EMC RecoverPoint Deploying RecoverPoint with SANTap Technical Notes
• EMC RecoverPoint Administrators Guide
• Cisco MDS 9000 Family Cookbook for SAN-OS
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 126 of 223
Part 1: Prepare to Migrate to SANTap Before implementing SANTap in a customer environment you must take steps to qualify the environment and plan the implementation (or kdriver migration). General Prerequisites (from the SANTap Deployment Tech Notes): To configure RecoverPoint to take advantage of SANTap services, the following is required: Notes: • Cisco Multilayer Director Switch with Storage Service Module and SANTap services installed at
the primary site. Cisco may decommission some of these models or replace them with others. Refer to Cisco’s publications for an updated product list.
• Hot migration to SANTap without application downtime is possible only if the system contains at least two SANTap switches supporting SANTap services.
• License for SANTap service for each Storage Services Module (SSM). • Port 23 between RPAs and the SAN switches must be available. • RecoverPoint software loaded on all RecoverPoint appliances (RPAs). • All relevant hosts, storage ports, and RPA ports must be connected via Fibre Channel to the same
VSAN. In a multi-switch environment, all connections must be made to the same VSAN on every switch.
• Hosts and RPAs must be zoned so that each can see the storage it must access. • Zoning must be planned according to the business needs of the organization. • RPAs can be zoned only after the RecoverPoint software has been loaded. Reference the RecoverPoint Release Notes and Support Matrix to verify code versions and reference the most current TS Practitioners Guide for SANTap implementation considerations.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 127 of 223
Step Action
1 Calculate the ITL count for your lab configuration for a single SSM. CVT ITLs = 1 per RPA node Discovery ITLs = 26 × HBA × RPA How many RPA nodes are in the cluster on this site? _______ CVT ITLs: __________ Discovery ITLs:__________ How many LUNs visible to your host will be replicated using SANTap? ________ How many HBAs from this host are attached to the switch? ___________ How many DVTs will be created? ________ Host ITLs: _____________ (#DVTs x #LUNs x #HBAs) Total ITLs: _____________ (Host ITLs + CVT ITLs + Discovery ITLs) Note: The DVT is correlated to the storage port where the LUNs are presented. In this dual path environment, the LUNs are visible via two different target ports. This will result is two DVTs. However, because this environment is a mirrored fabric the DVTs will be created on different SSMs. The ITL calculation is performed per SSM. BONUS: If your host is an IBM AIX system (requiring SCSI reservations) what would your ITL count be? EXTREME BONUS: If your host is part of a two node cluster and all LUNs require reservations what would your ITL count be?
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 128 of 223
Step Action 2 From the RecoverPoint Management Application, disable all your consistency groups.
Disable Consistency Groups
3 Detach all the volumes from the host based splitters (both sites).
Detaching Volumes
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 129 of 223
Step Action 4 Delete the host based splitters from the Management Console.
Delete Splitters
5 On your host, stop the kdriver. You can either disable the kdriver (so it does not restart if
the system is rebooted) or uninstall the kdriver.
Solaris Example: /etc/init.d/kdrv stop pkgrm kdriver
Windows Example
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 130 of 223
Step Action 6 Your objective is to migrate from kdriver to SANTap in a live environment. To do this,
work on one path at a time to ensure the host maintains access to the devices via the alternate path. From your host, use PowerPath commands to disable the HBA connected to the switch you are about to modify. Use the powermt display command to view status and identify the HBA #. Note: If this is your second switch on this site, perform the next step first to re-enable the path that was previously disabled (if not already done). This command is not available on Windows hosts so use PowerPath Monitor on Windows to disable the hba. Command Syntax powermt disable hba=<###> Command Example powermt disable hba=2304 Disable the Path
Note: After completing this step, proceed to Part 2 Deploying SANTap.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 131 of 223
Part 2: Deploying SANTap Step Action
1 Log in to the switch as admin and copy the running configuration to an ftp site. Command Syntax: copy run ftp://root@<IP Address>/<directory>/<filename> Important: Configuration changes on the switch should only be performed by one user at a time. Each workgroup should coordinate switch configurations to ensure only one user is changing a configuration. Before changing the switch configuration, each group should backup the existing configuration on their first assigned switch. In a customer environment you would perform a backup of all switches in the environment.
2 Configure the SSM for SANTap. Log into the switch as admin and use the show module command to identify the slot number of the SSM module. You will need this number to perform several of the SANTap configuration commands. This number may be different on each switch in your environment.
Command Example: mds# show module
Show Module – Output Example
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 132 of 223
Step Action 3 Verify the license using the show license command.
Command Example: mds# show license
Show License – Output Example
4 At the switch prompt, enter the configuration terminal using the conf t command. Then, enable the SANTap service on the SSM module using the ssm enable feature command. Command Syntax: mds(config)# ssm enable feature santap module <SSM#> Note: This step is performed once per switch. You do not need to run this step if SANTap was previously enabled by another lab group (see example below).
Failed Message When SANTap is Currently Enabled – Command Example
Note: You will need a front end VSAN for your host initiator (HBA) and a backend VSAN for your storage target and RPA entities. A VSAN containing the host, RPAs, and storage target should already exist from your kdriver implementation. Keep this VSAN and use it as your back end VSAN. You will need to create a new VSAN for the front end VSAN. You will create a DVT for the front end VSAN and move the host port from the back end to the front end VSAN. Follow the steps below.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 133 of 223
Step Action 5 At the switch prompt, enter the configuration terminal (if not already at the config prompt)
using the conf t command. Then enter the VSAN database using the command vsan database. Create the front end VSAN using the vsan command (see example below). Assign the VSAN a number and a name. To prevent conflicts with other lab groups, assign a number to your front end VSAN which is ten higher than your back end VSAN. For example, if you back end VSAN is 5 make your front end VSAN 15.
Command Syntax: mds# conf t mds(config)# vsan database mds(config-vsan-db)# vsan <FrontEndVSAN#> name <VSAN Name> mds(config-vsan-db)# exit
Creating the Front End VSAN – Command Examples
Note: In later steps you will create the DVT and the host to storage zone in the front end VSAN and move the host port into the front end VSAN.
6 From the config prompt, create the CVT in the back end VSAN. You will need to designate the SSM module number, your back end VSAN number, and designate a name for the CVT. Important: The CVT must be created in the back end VSAN. Do not create the CVT in the front end VSAN that you created in the previous step. Command Syntax: santap module <SSM#> appl-vsan <BackEndVSAN#> cvt-name <CVT Name> Command Example: mds(config)# santap module 2 appl-vsan 5 cvt-name group_5_CVT
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 134 of 223
Step Action 7 Verify the CVT was created correctly. You should see 10 Cisco PWWNs. Under the
“Type” column you should see 9 of the PWWNs listed as “init” (initiator) and 1 listed as “target”. In a later step you will need to add the CVT to a zone. You must include all 10 of the PWWNs.
Command Syntax: <do> show fcns database vsan <BackEndVSAN#>
Note: Add ‘do’ to the front of a show command if you are at the “config” prompt.
Command Example mds(config)# do show fcns database vsan 5
Finding the CVT in the Back End VSAN – Output Example
Set Persistent FCIDs (IBM AIX and HP-UX)
Setting persistent FCIDs must be done for IBM AIX and HP-UX. However, you can perform these steps for any platform. If your lab environment does not contain IBM AIX or HP-UX you may skip this section and proceed with the next section Create the DVT.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 135 of 223
Step Action 8 Determine the domain ID for the back end VSAN.
Command Syntax: show fcdomain domain-list vsan <BackEndVSAN#> Command Example:
mds# show fcdomain domain-list vsan 5
show fcdomain Output Example
Note: Use the decimal Domain ID number.
9 From the config prompt, set the domain ID for the front end VSAN to be identical to that of the back end VSAN.
Command Syntax: mds(config)# fcdomain domain <domainID> static vsan <FrontEndVSAN#> Command Example: mds(config)# fcdomain domain 8 static vsan 15
10 From the config prompt, restart the front end VSAN.
Command Syntax: mds(config)# fcdomain restart disruptive vsan <FrontEndVSAN#> Command Example: mds(config)# fcdomain restart disruptive vsan 15
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 136 of 223
Step Action 11 From the config prompt, enable the persistent FCID.
Command Syntax: mds(config)# fcdomain fcid persistent vsan <FrontEndVSAN#> Command Example: mds(config)# fcdomain fcid persistent vsan 15
Note: The switch should respond: FCID persistent feature is enabled
12 From the config-fcid-db prompt, make the FCID of the storage target port the same in the front end VSAN as in the back end VSAN. Use the show fcns database command to identify the Storage target port FCID.
Displaying the Storage Port FCID
Command Syntax: mds(config)# fcdomain fcid database mds(config-fcid-db)# vsan <FrontEndVSAN#> wwn <StoragePort WWN> fcid <StoragePort FCID> dynamic Command Example: mds(config)# fcdomain fcid database mds(config-fcid-db)# vsan 15 wwn 50:06:04:8a:ca:fe:64:13 fcid 0xa100c dynamic
13 Copy the running configuration to the startup configuration. Command Example: From the switch prompt run: copy run start From the config prompt run: do copy run start
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 137 of 223
Step Action 14 Create the DVT. Each physical storage target port requires a corresponding DVT. The
DVT and the storage target port will have the same PWWN (target-pwwn). The DVT resides in the front end VSAN (dvt-vsan) and the physical storage target port resides in the back end VSAN (target-vsan). In the lab configuration, the physical storage target port should already be in the back end VSAN. Creating the DVT is performed from the “config” prompt. If you are not already at the config prompt enter the configuration terminal using the conf t command. Command Syntax: santap module <SSM#> dvt target-pwwn <StoragePort WWN> target-vsan <BackEndVSAN#> dvt-name <name> dvt-vsan <FrontEndVSAN#> lun-size-handling 1 Command Example: mds(config)# santap module 2 dvt target-pwwn 50:06:01:60:10:60:1a:9c target-vsan 5 dvt-name Group_5_DVT dvt-vsan 15 lun-size-handling 1
15 Verify that the DVT was created properly. The command to display the DVT will show all DVTs configured. You may need to scroll through the output to find your DVT.
Command Syntax: <do> show santap module <SSN#> dvt
Command Example mds(config)# do show santap module 2 dvt
Display the DVTs Configured in the SSM – Output Example
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 138 of 223
Step Action 16 Copy the running configuration to the startup configuration.
Command Example: From the switch prompt run: copy run start From the config prompt run: do copy run start
Creating the Front End Zoneset and Moving the Host Port The VSAN you designated as the back-end VSAN should already contain a Host to Storage zone with the PWWNs of the host initiator port(s) and the storage target port(s). This zone must be recreated in the front end VSAN using the same PWWNs. The DVT and the physical storage target port have the same PWWN. The DVT is in the front end VSAN and the physical target is in the back end VSAN. The physical host initiator and the virtual host initiator (VI) also have the same PWWN. The physical host initiator will reside in the front end VSAN and the VI will be automatically created in the back end VSAN. Currently, the physical host initiator is logged into the back end VSAN. You will move it to the front end in a later step.
17 Display the active zone set in the back end VSAN. This will allow you to verify the PWWNs in the “Host to Storage” zone. You will need these PWWNs for the next step.
Command Syntax: <do> show zoneset active vsan <BackEndVSAN#>
Command Example mds(config)# do show zoneset active vsan 5
Host to Storage Zone in Back End VSAN – Output Example
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 139 of 223
Step Action 18 From the “config” prompt, create a new front end zone set and a “Host to Storage” zone that
matches the back end “Host to Storage” zone. Note: There are several ways to accomplish this task. The screen shot below provides an example of creating the zone set. The back end zone could also be copied to the front end using the command line or Fabric Manager GUI. Command Syntax: mds(config)# zoneset name <name> vsan <FrontEndVSAN#> mds(config-zoneset)# zone name <name> mds(config-zoneset-zone)# member pwwn <Storage Target PWWN> mds(config-zoneset-zone)# member pwwn <Host Initiator PWWN> mds(config-zoneset-zone)# exit mds(config-zoneset)#zoneset activate name <name> vsan <FrontEndVSAN#>
Creating and Activating the New Front End Zone Set – Command Examples
19 Copy the running configuration to the startup configuration.
Command Example: From the switch prompt run: copy run start From the config prompt run: do copy run start
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 140 of 223
Step Action 20 Use the show flogi database command to identify the interface number for the host initiator
port in the back end VSAN. You will need this number to move the host initiator to the front end VSAN.
Command Syntax: <do> show flogi database vsan <BackEndVSAN#>
Command Example mds# show flogi database vsan 5
Show Flogi Database – Output Example
21 From the “config-vsan-db” prompt, move the host initiator port (referencing the interface ID) from the back end VSAN to the front end VSAN. You only need to reference the front end VSAN number in the command. You do not need to reference the back end VSAN because the interface ID is unique.
Command Syntax: mds# config t mds(config)# vsan database mds(config-vsan-db)# vsan <FrontEndVSAN#> interface <interfaceID> switch(config-vsan-db)# exit
Moving the Host Interface ID to the Front End VSAN – Command Examples
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 141 of 223
Step Action Once the host initiator port is moved to the front end VSAN, it should log into the DVT
resulting in the virtual initiator being created in the back end VSAN. At this time, the host should be able to see its assigned devices. You can use inq or syminq on your host to verify the devices are visible. The following steps will run through a series of check to verify (from the switch perspective) that the configuration is correct to this point.
22 Verify that the host initiator port interface ID is listed in the flogi database for your front end VSAN.
Command Syntax: show flogi database vsan <FrontEndVSAN#>
Show Flogi Database – Command Example
23 Display the active zoneset in your front end VSAN. You should see that the host initiator PWWN has a “* fcid”. This indicates that the port has logged in and is active.
Command Syntax: show zoneset active vsan <FrontEndVSAN#>
Command Example: mds# show zoneset active vsan 15
Display Active Zoneset in Front End VSAN – Output Example
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 142 of 223
Step Action 24 Display the active zoneset in your back end VSAN. The VI PWWN should also have a “*
fcid”. Note: The FCID numbers should match if you configured persistent FCIDs on the switch. Command Syntax:
show zoneset active vsan <BackEndVSAN#> Command Example:
mds# show zoneset active vsan 5
Display Active Zoneset in Back End VSAN – Output Example
25 If the host initiator has successfully logged into the DVT and the VI was created in the back end VSAN, the DVT LUNs should now be visible. Viewing the DVT LUNs is similar to viewing the DVTs. The command will show all the DVT LUNs on the SSM module. You will have to scroll through the output to find your LUNs.
Command Syntax: <do> show santap module <SSM#> dvtlun
Command Example: mds# show santap module 2 dvtlun
DVT LUN Output Example
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 143 of 223
Part 3: Add RecoverPoint to the SANTap Environment Step Action
1 In a previous lab section, you created the CVT and displayed the ten CVT WWNs (show fcns database vsan <BackEndVSAN#>). To add the RecoverPoint Appliances to the SANTap configuration, create a zone in the back end VSAN that contains the ten (10) CVT WWNs and all RPAs port WWNs. Note: When creating the zone, you must add the CVT WWNs one line at a time. You will issue the command member pwwn 10 times for the CVT and once for each RPA port that is connected to this switch. Please see step 1a for commands reference.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 144 of 223
Step Action 1a Command Syntax:
mds(config)# zoneset name <name> vsan <BackEndVSAN#> mds(config-zoneset)# zone name <name> mds(config-zoneset-zone)# member pwwn <RPA5 PWWNa> mds(config-zoneset-zone)# member pwwn <RPA5 PWWNb> mds(config-zoneset-zone)# member pwwn <RPA6 PWWNa> mds(config-zoneset-zone)# member pwwn <RPA6 PWWNb> mds(config-zoneset-zone)# member pwwn <CVT PWWNs> mds(config-zoneset-zone)# exit mds(config-zoneset)#zoneset activate name <name> vsan <BackEndVSAN#> Command Examples: mds(config)# zoneset name Group_5_BackEnd30 vsan 5 mds(config-zoneset)# zone name RPA56_CVT mds(config-zoneset-zone)# member pwwn 50:01:24:82:00:8a:ea:50 mds(config-zoneset-zone)# member pwwn 50:01:24:82:00:aa:ea:51 mds(config-zoneset-zone)# member pwwn 50:01:24:82:00:9c:bc:40 mds(config-zoneset-zone)# member pwwn 50:01:24:82:00:cc:bc:41 mds(config-zoneset-zone)# member pwwn 24:ec:00:0d:ec:18:cb:42 mds(config-zoneset-zone)# member pwwn 23:ec:00:0d:ec:18:cb:40 mds(config-zoneset-zone)# member pwwn 23:ec:00:0d:ec:18:cb:41 mds(config-zoneset-zone)# member pwwn 23:ec:00:0d:ec:18:cb:42 mds(config-zoneset-zone)# member pwwn 23:ec:00:0d:ec:18:cb:43 mds(config-zoneset-zone)# member pwwn 23:ec:00:0d:ec:18:cb:44 mds(config-zoneset-zone)# member pwwn 23:ec:00:0d:ec:18:cb:45 mds(config-zoneset-zone)# member pwwn 23:ec:00:0d:ec:18:cb:46 mds(config-zoneset-zone)# member pwwn 23:ec:00:0d:ec:18:cb:47 mds(config-zoneset-zone)# member pwwn 23:ec:00:0d:ec:18:cb:48 mds(config-zoneset-zone)# exit mds(config-zoneset)# zoneset activate name Group_5_BackEnd30 vsan 5
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 145 of 223
Step Action 2 From the config prompt, create another new zone in the back end VSAN that contains all
RPAs port WWNs and the physical storage target port WWNs. Note: If you have migrated from a kdriver environment to a SANTap environment, this zone should already exist. You do not need to create another zone. Command Syntax: mds(config)# zoneset name <name> vsan <BackEndVSAN#> mds(config-zoneset)# zone name <name> mds(config-zoneset-zone)# member pwwn <RPA5 PWWNa> mds(config-zoneset-zone)# member pwwn <RPA5 PWWNb> mds(config-zoneset-zone)# member pwwn <RPA6 PWWNa> mds(config-zoneset-zone)# member pwwn <RPA6 PWWNb> mds(config-zoneset-zone)# member pwwn <Storage Target PWWN> mds(config-zoneset-zone)# exit mds(config-zoneset)# zoneset activate name <name> vsan <BackEndVSAN#> Command Examples: mds(config)# zoneset name Group_5_BackEnd30 vsan 5 mds(config-zoneset)# zone name RPA56_Sym345_fa4cb mds(config-zoneset-zone)# member pwwn 50:01:24:82:01:b3:bf:20 mds(config-zoneset-zone)# member pwwn 50:01:24:82:01:93:bf:21 mds(config-zoneset-zone)# member pwwn 50:01:24:82:01:82:cd:10 mds(config-zoneset-zone)# member pwwn 50:01:24:82:01:82:cd:11 mds(config-zoneset-zone)# member pwwn 50:06:04:8a:cb:1b:b8:63 mds(config-zoneset-zone)# exit mds(config-zoneset)# zoneset activate name Group_5_BackEnd30 vsan 5
3 Copy the running configuration to the startup configuration. Command Example: From the switch prompt run: copy run start From the config prompt run: do copy run start
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 146 of 223
Step Action 4 The CVT should be visible in the RecoverPoint Management Application as a valid splitter.
Log into the GUI, click on Splitters then click on the Rescan icon. Once the rescan is complete, click on the Add Splitter icon. Select the appropriate site and the CVT should be displayed as a splitter. Click OK to add the splitter. Note: If the CVT splitter is not visible, recheck the zoning created in the previous two steps. The next step (creating AVT zones) is dependant on the CVT being added as a splitter in this step. Do not proceed to the next step if you cannot complete this step.
Adding the CVT Splitter
Note: Perform this setup on all switches in the SANTap configuration. When running a CRR configuration with mirrored fabrics, you will have four CVT splitters when finished (two per site).
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 147 of 223
Step Action 5 From the RPA admin prompt, create the AVT zone for your RPA cluster. SSH into the
primary RPA node as “admin” and run the config_santap_avt_zones command. When prompted enter the user name for the Cisco switch (admin) and the password for that user. When prompted to “Enter switches” you can type in the full name of the CVT (including “Cisco –“) or press enter (once) to select all visible CVTs.
AVT Zone Creation – Command Example
Note: The RPA uses telnet to access the switch. In a customer environment, telnet may be disabled. Have it enabled for the purpose of building the AVT zone and it can be disabled again once you have finished.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 148 of 223
Step Action 6 A zone will be created in your back end VSAN but it will not be in the active zoneset.
Manually add the zone to the active zoneset. The zone name is: RPA-AVT-<CVT WWN> To display this zone you must view all zones in the VSAN (not just the active zoneset).
Command Syntax: show zone vsan <BackEndVSAN#>
Command Example: mds# show zone vsan 5
Note: The show zone command will display all the zones in the VSAN, including those in the active zoneset. You will need to scroll through the output to find the AVT zone. Be aware that the Cisco command line recognizes abbreviated commands. If you enter show zones (plural) rather than show zone (singular) it interprets the command as show zoneset which results in your AVT zones not being displayed.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 149 of 223
Step Action 7 From the config prompt:
a) add the AVT zone to the active zoneset b) add the RPA ports to the AVT zone (if they are not already there! – in this case skip
this step and move to step c) c) activate the zoneset
Command Syntax: mds(config)# zoneset name <name> vsan <BackEndVSAN#> mds(config-zoneset)# member <AVT zone name> mds(config-zoneset)# zone name <AVT zone name> mds(config-zoneset-zone)# member pwwn <RPA5 PWWNa> mds(config-zoneset-zone)# member pwwn <RPA5 PWWNb> mds(config-zoneset-zone)# member pwwn <RPA6 PWWNa> mds(config-zoneset-zone)# member pwwn <RPA6 PWWNb> mds(config-zoneset-zone)# exit mds(config-zoneset)# zoneset activate name <name> vsan <BackEndVSAN#> Command Examples: mds(config)# zoneset name Group_5_BackEnd30 vsan 5 mds(config-zoneset)# member RPA-AVT-2584000dec18d4c2 mds(config-zoneset)# zone name RPA-AVT-2584000dec18d4c2 mds(config-zoneset-zone)# member pwwn 50:01:24:82:00:8a:f3:4e mds(config-zoneset-zone)# member pwwn 50:01:24:82:00:aa:f3:4f mds(config-zoneset-zone)# member pwwn 50:01:24:82:00:7b:e1:3a mds(config-zoneset-zone)# member pwwn 50:01:24:82:00:bb:e1:3b mds(config-zoneset-zone)# exit mds(config-zoneset)# zoneset activate name Group_5_BackEnd30 vsan 5
8 Copy the running configuration to the startup configuration.
Command Example: From the switch prompt run: copy run start From the config prompt run: do copy run start
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 150 of 223
Step Action 9 From your host use PowerPath commands to enable the HBA connected to the switch you
just finished configuring. Use the powermt display command to view status and identify the HBA #. Note: If this is your first switch on this site, skip this step and proceed to Part 2 “Deploy MDS SANTap”.
Command Syntax powermt enable hba=<###>
Command Example powermt enable hba=2304
Enable Path
10 Repeat the procedure to configure SANTap on the next switch. If you have completed both switches on both sites, proceed to the next section to restart replication.
End of lab exercise.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 151 of 223
Part 4: Restart Replication: SANTap Implementation Complete for all Switches
Step Action
1 After completing the SANTap implementation on all four switches, restore Consistency Groups to an operational state. From the RecoverPoint Management Console, highlight a splitter and open Splitter Properties. Attach the replication volumes to all splitters (on each site).
Attach Volumes To Splitters
2 From the RecoverPoint Management Console, re-enable the consistency group. Monitor
the consistency group to ensure replication resumes normally.
Enable Group
End of Lab Exercise
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 152 of 223
Lab Exercise 9: Migrate to Brocade Fabric Splitter
Purpose:
Migrate an existing host based splitter implementation to a Brocade fabric based splitter. Install the RecoverPoint driver onto the Connectrix AP-7600B, and verify correct version of Fabric O/S (FOS) and SAS package are installed. Upgrade FOS and SAS if necessary. Perform zoning and masking.
Tasks: Prepare the AP-7600B for use as a RecoverPoint Intelligent Fabric Splitter • Verify Correct version of FOS is installed and update if necessary • Verify Correct version of SAS is installed and update if necessary • Install RecoverPoint Driver and verify it is operational • Create zones needed to support fabric splitting in Frame Redirect • Perform LUN masking needed to support fabric splitting Frame
Redirect mode
References: Deploying RecoverPoint with the Brocade AP7600 and FA4-18 Technical Notes
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 153 of 223
Part 1: Prepare to Migrate to Brocade SAS Step Action
1 From the RecoverPoint Management Application, disable all your consistency groups.
Disable Consistency Groups
2 Detach all the volumes from the host based splitters (both sites).
Detaching Volumes
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 154 of 223
Step Action 3 Delete the host based splitters from the Management Console.
Delete Splitters
4 On your host, stop the kdriver. You can either disable the kdriver (so it does not restart if
the system is rebooted) or uninstall the kdriver.
Solaris Example: /etc/init.d/kdrv stop pkgrm kdriver
Windows Example
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 155 of 223
Step Action 5 Your objective is to migrate from kdriver to Brocade in a live environment. To do this, you
work on one path at a time to ensure the host maintains access to the devices via the alternate path. From your host, use PowerPath commands to disable the HBA connected to the switch you are about to modify. Use the powermt display command to view status and identify the HBA #. Note: If this is your second switch on this site, perform the next step first to re-enable the path that was previously disabled (if not already done). This command is not available on Windows hosts so use PowerPath Monitor on Windows to disable the hba. Command Syntax powermt disable hba=<###> Command Example powermt disable hba=2304
Disable the Path
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 156 of 223
Part 2: Verify AP-7600B is prepared for RecoverPoint Installation
Step Action
1 Telnet as admin into your assigned RecoverPoint AP7600B.
Telnet to Switch
2 Display the firmware revision on the switch using the firmwareshow or version commands.
Output: firmwareshow
3 If your switch is running the FOS version specified by the instructor, skip to Step 6. To upgrade FOS use the firmwaredownload CLI command (or use Web Tools). Enter FOS as the Type of Firmware. Enter the IP address and user name for the system where the file is located. Enter the full directory path to the file. Select FTP as the Network Protocol. Enter the system password when prompted.
4 Once the download is completed, the switch reboots.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 157 of 223
Step Action 5 Once the switch reboots, log back in and monitor the firmware installation using the
firmwaredownloadstatus command.
6 Once the FOS image is updated, verify the version of SAS on your switch using the firmwareshow command. If your switch is already running the version of SAS specified by the instructor, skip to Step 9.
7 Use the firmwaredownload command to install the correct version of SAS. Enter SAS as the Type of Firmware. Enter the IP address and user name for the system where the file is located. Enter the full directory path to the file. Select FTP as the Network Protocol. Enter the system password when prompted.
8 Continue to issue the firmwaredownloadstatus commands until you see that the
SAS image loaded successfully.
9 Use the ipaddrshow command to verify that the eth0 port on the blade processor is configured with the IP address given to you by your instructor.
Output: ipaddrshow
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 158 of 223
Step Action 10 Telnet into the Blade Processor (BP) IP address you set up in the previous step. Login as
root user, with the default password given to you by your instructor.
Telnet to BP
11 Change directory to /tmp and open a FTP connection to the system where the RecoverPoint driver software is located.
FTP from Blade Processor
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 159 of 223
Step Action 12 Change directory to the location of the RecoverPoint driver and download the file to the
switch BP.
FTP RecoverPoint Driver
13 Use the ls –l command to verify the file was downloaded successfully.
Output: ls –l
14 Make the file executable with the chmod command. Command Syntax: chmod +x <filename> Command Example: chmod +x
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 160 of 223
Step Action 15 Execute the RecoverPoint installation program. When prompted, enter a host name. It is
asking you for a unique name for the blade processor. The name you give has the Blade Processor IP address appended to it, and that becomes the name of the splitter when you add it to RecoverPoint.
Installing the RecoverPoint Driver
16 When the installation is complete, reboot the blade processor. Reboot the Blade Processor
17 After the BP reboots log back in and verify that the RecoverPoint driver is running . This can only be seen from the blade processor when you login in as root. Change directory to /thirdparty/recoverpoint/install using the cd command. Use the kdrv status and ps –A commands to view the status.
Command Output: kdrv status & ps -A
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 161 of 223
Part 3: Verify the SysVI exists and check fabric logins
Step Action
1 Once the RecoverPoint driver is started, the switch creates the System VI. This can be seen with the nsshow command on the AP-7600B. You need the WWN of the System VI to perform zoning in a Multi-VI configuration. If you are implementing Frame Redirect, the System VI WWN is added to the correct zone automatically after running the zone create script.
Output: nsshow
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 162 of 223
Step Action 2 Details on fabric wide logins can be seen with the nscamshow command on the AP-
7600B. This information helps you to determine the PWWNs of the components logged into a remote (ISLed) switch. This does not apply to our lab environment as we have only a single AP7600B. Output: nscamshow
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 163 of 223
Part 4: Configure Zoning – Frame Redirect
The procedure in this section is for Frame Redirect mode. If you are going to implement Multi-VI mode, go to Lab S6 – Configure Multi-VI – in the supplement lab guide. Step Action
1 To configure Frame Redirect, set up zoning for the RPAs. Reference the table below for the list of zones needed for Frame Redirect. In this exercise, you are migrating from a host splitter environment. Delete the RPA ports to Host initiator zone when you are finished implementing Frame Redirect.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 164 of 223
Step Action 2 Run the cfgacvtshow command to display the current effective configuration on your
switch and verify which zones are part of the configuration. Your zones should be configured for host based splitting. The existing RPA ports to storage target zone is the RPA Initiator Zone. If this zone(s) does not exist, create it now. Make note of the configuration name, you need this information to add zones to the configuration in a later step.
3 Create the RPA Target Zone and RPA Front-end Zone. On the BP, run the script: /thirdparty/recoverpoint/install/zoning_script.sh The script creates the RPA target zone (HOSTNAME_FR_RPA_Target_Zone) and the RPA front-end zone (HOSTNAME_RPA_Front_End_Zone), and adds the system VI and all possible VI WWNs to the RPA target zone, and the appliance virtual targets (AVTs) to the RPA front-end zone.
Example Running Zoning Script
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 165 of 223
Step Action 4 Add all RPA ports to both zones.
Command Syntax: zoneadd “RPA Target Zone”, “RPA1 PWWNa; RPA1 PWWNb; RPA2 PWWNa; RPA2 PWWNb” zoneadd “RPA Front-end Zone”, “RPA1 PWWNa; RPA1 PWWNb; RPA2 PWWNa; RPA2 PWWNb”
5 Add the new zones to the configuration.
Command Syntax: cfgadd “configuration name”, “zone name; zone name”
Adding Zones
6 Enable the configuration. This makes the splitter visible in the RecoverPoint Management Application which you add in the next step. Command Syntax: cfgenable “configuration name”
Enabling Configuration
7 Use the RecoverPoint Management Application or CLI to add the Connectrix-based splitter to the RecoverPoint configuration.
Adding the Splitter in the GUI
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 166 of 223
Step Action 8 SSH into the RPA as admin and run the bind_host_initiators command to configure host
binding to storage targets. Enter the site name and the splitter name when prompted. Choose 1 for Yes to enable Frame Redirect. Select your physical host initiator PWWN from the list. Enter the PWWN of the physical storage target this host initiator must see. If the splitter was not successfully added in the previous step, the bind operation will not work.
Example: bind_host_initiators
This command creates the virtual initiators and virtual targets that correspond to the host initiators and storage targets involved, and that enable the frame redirect mechanism to direct the data frames from the initiator to the target through the virtual entities. Each initiator should be bound to all targets that expose protected LUNs to it. You can bind multiple initiators to multiple targets.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 167 of 223
Step Action 9 The bind command will not return any virtual WWNs. You do not need these WWNs
because the binding relationship is managed by the name server on the switch. You can view the binding relationship between the physical host and physical storage target from the RPA admin prompt by running the get_initiator_bindings command.
Output: get_initiator_bindings
10 If you are deploying RecoverPoint non-disruptively, enable the paths that you disabled at the beginning of this exercise. Verify that host initiators can see the virtual targets and the correct LUNs in them, and that I/Os continue properly. Verify that all paths over all switches are working.
11 Currently, there should be one zone in your active configuration which is no longer needed. The zone containing the RPA ports and the host initiator PWWN (which was used for the host based splitter configuration) should be removed from the active configuration. Command Syntax: cfgremove “configuration name”, “zone name” cfgenable “configuration name”
12 Repeat these steps in the second fabric as well as at site 2 (both fabrics).
End of Lab Exercise
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 168 of 223
This page intentionally left blank.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 169 of 223
Lab Exercise 10: Working With CLR
Purpose:
Add a CDP copy to an existing CRR consistency group. Work with CDP and CRR images for logged access, failover, and recovery.
Tasks: In this lab, you perform the following tasks: • Add CDP copy to CRR consistency group • Perform CDP failover/failback • Perform CDP restore operation • Perform CRR failover, switching production to remote site
References: • Student Guide • EMC RecoverPoint Administrators Guide
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 170 of 223
Part 1: Add CDP Copy to CRR Consistency Group Step Action
1 Add a CDP copy to your consistency group by either right-clicking your consistency group in the three panel and selecting Add Copy or clicking on the Add Copy icon.
Groups Menu
1a Name the new local copy and leave the remaining settings at the defaults. Click Next.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 171 of 223
Step Action 2 Expand the Replication Sets, then highlight the CDP copy and click Add Volume.
New Copy Wizard
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 172 of 223
Step Action 3 Select a volume to use as the CDP replica and click OK.
Select Volume
4 Under Journals, highlight the CDP copy name and click Add New Journal Volume.
New Copy Wizard – Journals
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 173 of 223
Step Action 5 Select a volume to be used as a journal for the CDP copy and click OK.
Select Volume
6 If the procedure is complete, the Finish button is now enabled. Click Finish to stop adding
the replication and journal volumes to the CDP copy.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 174 of 223
Step Action 7 Right click on the CDP copy and select Enable Copy to start the data transfer.
Enable CDP Copy
8 Click Yes to confirm and star data transfer.
9 View the consistency group status. The production volume appears in the middle with the
CDP copy to the left (Local Replica) and the CRR copy to the right (Remote Replica). Consistency Group Status - Example
End of Lab Exercise
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 175 of 223
Part 2: CDP Failover Step Action
1 Umount the production copy volumes at the local site (site 1) using either kutils or mountvol.
2 Create a new bookmark image.
Create Bookmark
3 Access the new bookmark from the dropdown menu under the CDP volume (Local Copy).
Note that the bookmark can be selected under either copy.
Select Bookmark Image
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 176 of 223
Step Action 4 Once the bookmark is fully distributed, select Failover from the dropdown menu under the
CDP copy.
Failover CDP
5 Click Yes to confirm the failover.
Confirm Failover
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 177 of 223
Step Action 6 Click Yes on the warning message. Note that when a failover is performed from the CDP
copy, the data transfer to the remote replica is paused.
Warning Message
7 View the status of the consistency group. Notice that the CDP copy now has the Local
Source role.
Consistency Group Status
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 178 of 223
Step Action 8 Set the CDP copy to be the production volume by clicking the dropdown menu and
selecting Set <name> as Production.
Set CDP Copy as Production
9 Click Yes to confirm.
Confirmation Message
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 179 of 223
Step Action 10 Click Yes on the warning message. This requires a full sweep to the remote replica.
Warning Message
11 View the status of the consistency group. Notice the CDP copy is now the Production
Source in the center with the former production volume acting as the local replica.
Consistency Group Status
12 Repeat this process to return the CDP copy to the role of Local Replica.
End of Lab Exercise
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 180 of 223
Part 3: CDP Restore Step Action
1 Using either kutils or mountvol umount the production replication volumes at the local site (site 1).
2 Create a new bookmark.
Create Bookmark
3 Access the new bookmark from the local copy (CDP) dropdown menu.
Access Bookmark Image
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 181 of 223
Step Action 4 Once the bookmark is fully distributed, select Recover Production from the dropdown
menu under the local replica.
Recover Production
5 Click Yes on the confirmation message.
Confirmation Message
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 182 of 223
Step Action 6 Click Yes on the warning message.
Warning Message
7 View the status of the consistency group. Notice the production copy is being restored and
the remote replication is paused.
Consistency Group Status
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 183 of 223
Step Action 8 Create a new bookmark.
Create Bookmark
9 Access the new bookmark from the production volume dropdown menu.
Access Bookmark
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 184 of 223
Step Action 10 Once the bookmark is fully distributed, resume production on the production volume by
clicking the Resume Production icon.
Resume Production
11 Click Yes to confirm the resume production operation.
Confirmation Message
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 185 of 223
Step Action 12 Click Yes on the warning message. Notice the journal associated with the production
volume will be cleared.
Warning Message
13 View the status of the consistency group.
Consistency Group Status
End of Lab Exercise
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 186 of 223
Part 4: Perform CRR Failover with a CLR Consistency Group Step Action
1 Using either kutils or mountvol umount the production replication volumes at the local site (site 1).
2 Create a new bookmark.
Create Bookmark
3 Access the new bookmark from the remote copy (CRR) dropdown menu.
Access Bookmark
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 187 of 223
Step Action 4 Once the bookmark is fully distributed perform a failover by selecting Failover from the
remote replica dropdown.
Perform Failover
5 Click Yes to confirm the failover operation.
Confirmation Message
6 Click Yes on the warning message. Notice the CDP replication will be paused.
Warning Message
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 188 of 223
Step Action 7 View the status of the consistency group. The role of the remote replica is now Remote
Source and the CDP data transfer is paused.
Consistency Group Status
8 Set the remote copy as the production copy by clicking the dropdown menu under the
remote volume and selecting Set <name> as Production.
Setting Replica as Production
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 189 of 223
Step Action 9 Choose a local copy to remove from the configuration. When setting the remote copy as the
production, you cannot keep two remote copies.
Choose Copy to Remove
10 Click Yes on the warning message.
Warning Message
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 190 of 223
Step Action 11 View the status of the consistency group. The role of the remote copy is Production Source
and the second local copy (CDP) is removed.
Consistency Group Status
12 Perform the necessary steps to failback the consistency group resulting in the local copy
having the role of the Production Source and the remote copy the role of Remote Replica.
End of Lab Exercise
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 191 of 223
This page intentionally left blank.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 192 of 223
Lab Exercise 11: Upgrade RPAs
Purpose:
Utilize the ISO download feature to perform an RPA software upgrade.
Tasks: This lab requires performing the following activities:
• Disable consistency groups • Detach RPA from cluster. • Load upgrade software and reboot • Reapply settings • Select Repository • Enable consistency groups
References: • Lecture material
• Installation Guide
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 193 of 223
Part 1: Prepare Consistency Groups Note: Each RPA must be updated individually. After updating an RPA, for maximum efficiency immediately update its peer on the other side of the WAN (at the other site); for instance, RPA 1 Site 1, RPA 1, Site 2, etc. Step Action
1 Check in the GUI that all RPAs are functioning correctly. If there are any errors, resolve the issue before continuing.
Check RPA Status
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 194 of 223
Step Action 2 In a two (or more) node cluster, manually move all CGs to an RPA that is not being
upgraded. Once the upgrade for RPA 1 (on both sites) is completed, move all CGs back to RPA1 and upgrade RPA 2 on both sites. This allows replication to continue during the upgrade.
Change Preferred RPA
Note: If you are working in a lab environment configured with a single node cluster, disable your consistency groups.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 195 of 223
Part 2: Download ISO Image Step Action
1 Log into RPA 1 on site 1 as boxmgmt and detach the RPA from the cluster. From the Main menu select Cluster Operations, then Detach from cluster.
Detach from Cluster
** Main Menu ** [1] Installation [2] Setup [3] Diagnostics [4] Cluster operations [5] Shutdown / Reboot operations [Q] Quit pod1 RPA1: 4 ** Cluster operations ** [1] Attach to cluster [2] Detach from cluster [B] Back [Q] Quit pod1 RPA1: 2 Detach from cluster succeeded.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 196 of 223
Step Action 2 Enter b to go back to the Main menu. From the Main menu, select Installation, then
Upgrade Wizard. From the Upgrade Wizard menu, select Download ISO.
Upgrade Wizard
3 Enter the IP address of the lab system where the ISO image is located. Accept the default FTP port and enter the user name and password for the system when prompted. You must provide the full path to the ISO image (remember that UNIX is case sensitive) and the complete filename of the ISO image.
Download ISO Image
Enter the name of the ftp server you wish to download ISO file from: 10.127.14.60 Enter the port number to connect on the FTP server (0 to 65535) (Default '21'): 21 Enter the FTP user name:root Enter the FTP password:******** Enter the location of the ISO file on the FTP server:/software/RP/3.2/rel3.2.SP1_g.63 Enter the name of the ISO file on the FTP server:rel3.2.SP1_g.63_release_emc_md5_717ae86eeb180448e6dbdb671e7c4.iso
** Main Menu ** [1] Installation [2] Setup [3] Diagnostics [4] Cluster operations [5] Shutdown / Reboot operations [Q] Quit pod1 RPA1: 1 ** Installation ** [1] Use wizard to enter setup information [2] Import setup from another RPA [3] Upgrade [B] Back [Q] Quit pod1 RPA1:3 ** Upgrade ** [1] Download ISO [2] Apply downloaded ISO [3] Auto Upgrade [B] Back [Q] Quit pod1 RPA1:
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 197 of 223
Step Action 4 Once the file is downloaded, enter y to start the upgrade.
Begin Upgrade
5 When the upgrade completes, press Enter to reboot the RPA.
Reboot RPA
Upgrade succeeded, RPA is going to reboot Press ENTER to continue...
Downloading file… The file had been downloaded successfully Calculating checksum ISO checksum verified successfully Do you want to perform ISO upgrade now? (y/n) >>y Upgrading ISO... Caution: Do not use Ctrl+C during ISO upgrade, as it may cause unrecoverable RPA errors.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 198 of 223
Part 3: Reapply Settings Step Action
1 Log into the RPA as boxmgmt and redefine the site layout. Do not re-configure the management IP address. Enter number of sites (2) and the number of RPAs on site (2). Do not enable replication over fibre channel.
Define Site Layout
2 From the Main menu, select Setup then View. You see that there are no settings on the
upgraded RPA.
View Settings
|Parameter |Site 1 |Site 2 | |----------------------------|-----------------------|-----------------------| | Site Name | pod1 | pod2 | | Site Management IPv4 | 10.127.14.29 | 10.127.14.149 | | Management Default Gateway | 10.127.14.1 | 10.127.14.254 | | IPv4 | | | | Management Subnet Mask IPv | 255.255.255.128 | 255.255.255.128 | | 4 | | | | Site Management IPv6 | | | | Management Default Gateway | | | | IPv6 | | | | Management Subnet Mask IPv | | | | 6 | | | | WANDefault Gateway | 10.127.14.1 | 10.127.14.254 | | WANSubnet Mask | 255.255.255.128 | 255.255.255.128 | | Time Zone | Eastern Standard Time | Eastern Standard Time | | Primary DNS Server | 10.127.70.1 | 10.127.70.1 | | Secondary DNS Server | 10.127.70.2 | 10.127.70.2 | | Local Domain | | | | Number of Virtual Ports | N/A | N/A | | Initiator Only Mode | No | No | | Number of Exposed LUNs | N/A | N/A | Press enter to continue, '>>' to display entire list and Ctrl-C to exit the display
Linux RecoverPoint_pod1_RPA1 2.6.27.6-k100.x64.deb #24 SMP Mon Apr 6 01:27:58 EDT 2009 x86_64 Last login: Tue Sep 8 18:00:02 2009 from 10.127.118.6 Loading RPA settings... done Initializing Installation Manager... done Installation Manager - RecoverPoint Version 3.2(e.146) Site 1 RPA 1 ** Main Menu ** [1] Installation [2] Setup [3] Diagnostics [4] Cluster operations [5] Shutdown / Reboot operations [Q] Quit pod1 RPA1:
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 199 of 223
Step Action 3 From the Setup menu, select Get remote settings. Select the WAN interface. In the lab
environment ,you do not need to designate a temporary IP address or gateway. When prompted, enter the IP address of the remote RPA.
Get Settings from Remote RPA
4 Once the settings have been copied from the remote RPA select Apply from the Setup menu.
Apply Settings
** Setup ** [1] Modify [2] Configure repository volume [3] Get remote settings [4] Retrieve previous settings [5] Apply [6] View [7] Reset settings changes [8] Advanced options [9] Console Configuration [B] Back [Q] Quit pod1 RPA1: 3 . . . The current Site is pod1 and the RPA is 1 Would you like to apply those settings now (y/n)?y
** Setup ** [1] Modify [2] Configure repository volume [3] Get remote settings [4] Retrieve previous settings [5] Apply [6] View [7] Reset settings changes [8] Advanced options [9] Console Configuration [B] Back [Q] Quit pod1 RPA1: 3 1. Management interface 2. WAN Through which interface do you want to get the settings (1 to 2): Through which interface do you want to get the settings (1 to 2): 2 Do you want to configure a temporary IP address (y/n)? n Enter the IP address for the RPA from which you want to import the settings:
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 200 of 223
Step Action 5 From the Setup menu, select Configure repository volume. From the Select Method
menu, choose Select an existing repository volume. From version 3.1 up you do not need to reformat the repository for this upgrade.
Configure Repository Volume
6 The only volume displayed should be the volume that was previously formatted as the
repository. Select the volume and press enter.
Select Repository Volume
Index: 1 Size: 3.0GB Vendor: DGC Product: RAID 5 Name: Repository (0) UID: 60,06,01,60,4e,c0,20,00,4c,75,bd,41,0e,8d,de,11 Site: pod1 Volume Paths: |---|--------------------|------|----------------|----| |Prt|WWN |Ctrl |Serial |LUN | |---|--------------------|------|----------------|----| | 0 | 0x500601603ce02e6f | SP-A | APM00091501663 | 10 | | 3 | 0x500601603ce02e6f | SP-A | APM00091501663 | 10 | | 1 | 0x5006016b3ce02e6f | SP-B | APM00091501663 | 10 | | 2 | 0x5006016b3ce02e6f | SP-B | APM00091501663 | 10 | |---|--------------------|------|----------------|----| Enter your selection (1 to 1):
** Setup ** [1] Modify [2] Configure repository volume [3] Get remote settings [4] Retrieve previous settings [5] Apply [6] View [7] Reset settings changes [8] Advanced options [9] Console Configuration [B] Back [Q] Quit pod1 RPA1: 2 1. Select an existing repository volume 2. Format a volume as a repository volume Select the operation for the volume (1 to 2):1
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 201 of 223
Step Action 7 Enter b to go back to the Main menu. From the Main menu, select Cluster operations,
then Attach to cluster. When prompted, enter y to confirm.
Attach to Cluster
8 Verify the RPA is up and running then repeat the ISO upgrade procedure for the corresponding RPA on site 2. Then repeat for the remaining RPAs in each site.
** Cluster operations ** [1] Attach to cluster [2] Detach from cluster [B] Back [Q] Quit pod1 RPA1: 1 Do you want to attach the RPA to the cluster (Note: RPA will be rebooted) (y/n)? y You have configured an NTP server for the RPA cluster. There is no need to set the system clock. Do you want to do so anyway? (y/n)? n RPA is going to be attached and reboot. Attach to cluster succeeded
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 202 of 223
Part 4: Working with RecoverPoint 3.2 NDU Tool Step Action
1 The upgrade to 3.2 involves using the Upgrade Tool, a Java-based wizard, which allows you to perform a non-disruptive upgrade of RPA code from 3.1 up. The tool is composed of two flows - Prepare Upgrade and Apply. The Prepare Upgrade Wizard guides you through the verification of RecoverPoint credentials and downloads ISO files to all RPAs in the environment. The Apply step applies the upgrade to all RPAs. The upgrade procedure does not upgrade RecoverPoint splitters (KDriver, CLARiiON, SANTap, and Brocade). However, splitters compatible with RecoverPoint 3.1 and higher will continue to work with RecoverPoint 3.2. As well, a RecoverPoint 3.2 environment can simultaneously support 3.1 and higher splitters and 3.2 splitters. Note1: Upgrade to 3.2 splitters only after upgrading RPAs to 3.2 code. Note2: Java version 6 update 7 (1.6.0_7) or higher must be installed on the machine you are performing the upgrade from. To verify the version, run “cmd.exe” and type “java –version”. If needed, download and install Java before proceeding.
2 For the sake of lab activities you only will run the Prepare flow of the NDU tool. The apply process cannot be performed because the RPAs in the lab already have the latest code version running on them.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 203 of 223
Step Action 3 From Software Downloads section of Powerlink, download the file
RecoverPoint_3.2_Upgrade_Tool_for_Windows.zip to your management station.
4 Extract (unzip) the file contents.
5 Move to the “ndu” / “3.2 SP1 GUI” folder and run “ndutool.exe”. Review the conditions for upgrading your environment and verify that you have fulfilled them.
6 Once the tool is executed, a Wizard is displayed. Choose the Prepare Flow wizard.
7 Enter login credentials: • Enter a site management or box IP address. • Enter installation username and password (default boxmgmt). • Enter CLI username and password (default admin).
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 204 of 223
Step Action 8 Enter FTP credentials of FTP server where RecoverPoint 3.2 ISO image is located:
• Either for one or two sites. • FTP IP address. • Port number to connect on FTP server (default is 21). • FTP username. • FTP password. • Location of ISO file on FTP server. • Name of the ISO file on FTP server.
9 Download ISO file to all RPAs.
10 Save upgrade configuration file: • Enter location to save file (default is C:\Documents and Settings\<username>\). • Enter name of file (default is ndu_input.properties).
Press <Enter> to accept defaults. You will also be prompted that modifications may exist in the system you are upgrading, and if found, they will be saved to a file. The message will also say that if you are upgrading from 3.1 or 3.1 SP1, you must follow the procedure in “Verify and collect system modifications files” on the Installation Guide and press n (no) in the following question.
11 When asked to apply the upgrade now enter ‘n’.
The apply process cannot be performed because the RPAs in the lab already have the latest code version running on them. This exercise reinforces the knowledge obtained in the class regarding the use of the RecoverPoint 3.2 NDU Tool.
End of Lab Exercise
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 205 of 223
Lab Exercise 12: Troubleshooting
Purpose:
In this lab, you troubleshoot RPA installations, operations, and replications. You also implement remote support; SNMP & SMTP reporting; log file generation, etc.
Tasks: In this lab, you perform the following tasks: • View details and logs via the CLI • View details and logs via the GUI • Backup and restore of an Appliance’s meta data with CLI • Configure remote support, SNMP & SMTP reporting
References: • Lecture material
• Administrator’s Guide
• CLI Guide
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 206 of 223
Part 1: Saving and Restoring RPA Settings Step Action
1 Backup the RecoverPoint repository volume using ssh from a UNIX server. Connect to the appliance as “admin” and run the command save_settings. To save the output from save_settings, simply redirect the output of STDOUT to a file on the local UNIX host. Example: ssh [email protected] save_settings > backup.file
Taking Backup with “save_settings”
2 View the contents of the backup file. Example: more backup.file
Viewing the Backup File s1-sun2 /rpa more s1spa1_repo_backup ###### save_settings output (version: 3.2 (e.131)) ######; clear_settings –f; rescan_san volumes=none; ############################## # Settings for prod_cg_1 ############################## create_group name=”prod_cg_1” preferred_box=”RPA2” host_os=”solaris”; edit_group_settings group=”prod_cg_1” create_copy name=”prod_copy_1” group=”prod_cg_1” site=”s1-bos2”; create_copy name=”prod_1_remote” group=”prod_cg_1” site=”s2-lilrock2”; create copy name=”prod copy 2” group=”prod cg 1” site=”s1-bos2”;
s1-sun2 /rpa ssh [email protected] save_settings > s1spa1_repo_backup Password: s1-sun2 /rpa ls s1rpa1_repo_backup s1-sun2 / rpa
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 207 of 223
Step Action 3 Restore the RPA settings from the UNIX host where the backup file is located. Use ssh to
connect as admin and redirect the contents of the backup file as input to the command. Example: ssh [email protected] < backup.file
Restoring RPA Settings
End of Lab Exercise
s1-sun2 /rpa ssh [email protected] < s1spa1_repo_backup Pseudo-terminal will not be allocated because stdin is not a terminal. Password: Settings cleared successfully. Rescanning splitters in s1-bos SAN... Splitters successfully rescanned Rescanning splitter view for s1-sun2-ora2 in s1-bos2... Rescan successful Rescanning splitters in s2-lilrock2 SAN... Splitters successfully rescanned Rescanning splitter view for s1-sun2-ora2 in s2-lilrock2... Rescan successful Group prod_cg_1 created successfully. Request to edit group registered successfully. Copy prod_copy_1 created successfully. Copy prod_1_remote created successfully. Copy prod_copy_2 created successfully. Replication set prod_1_replica_set created successfully. Production copy set successfully Volume added successfully Volume added successfully
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 208 of 223
Part 2: Implementing SNMP Traps / Email Alert / Syslog Reporting
Step Action
1 From the System Menu, select SNMP Settings.
Input Details to Configure SNMP Trap Information
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 209 of 223
Step Action 2 Select the System menu then the Syslog Settings menu option. Input the criteria to allow
communication to your preferred syslog server.
Syslog Settings
3 From the System Menu, select Alert Settings. Click the checkbox for Email System
Enabled. Enter a relay compatible SMTP server name/IP address. Enter a valid sender address. Click the checkbox for the desired report(s). Add a rule for what type of alert to send via email.
Alert Settings
End of Lab Exercise
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 210 of 223
Part 3: Troubleshooting Using the RecoverPoint Management Application
Step Action
1 If your current status is all OK in the GUI, you can create a status change by detaching one of the RPAs from the cluster. Connect to one of your RPAs using ssh (or putty) as boxmgmt. Example: ssh [email protected]
2 From the Main Menu, select Cluster Operations > Detach From Cluster and confirm y when prompted.
Detaching RPA From Cluster
** Main Menu ** [1] Installation [2] Setup [3] Diagnostics [4] Cluster operations [5] Shutdown / Reboot operations [Q] Quit pod1 RPA1: 4 ** Cluster operations ** [1] Attach to cluster [2] Detach from cluster [B] Back [Q] Quit pod1 RPA1: 2 Detach from cluster succeeded.
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 211 of 223
Step Action 3 The GUI will list unknowns as a ? and errors as !. If you detach an RPA from the cluster, the
status for the RPA should change to Error. Viewing these preliminary error messages determines what to investigate further (system configuration, logs, SAN Configuration, Networking, repository configuration, etc).
Status Overview
4 After review status information provided by the GUI, examine the system logs by highlighting
Logs.
Logs Tab
5 Highlight an error message and click the Log Event Properties icon to view more details.
Log Details
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 212 of 223
Step Action 6 Utilize the Filter Log feature to examine an issue more closely. This allows you to determine if
there have been a series of events related to this component. In the example, the Topics are reduced to just RPA, the Scope is set to Detailed, the Level is set to Error and a text filter of RPA 1 is used.
Modifying Logs View
7 View the Logs tab after changing the log filter.
Filtered Log View
8 If you detached the RPA from the cluster in Step 1, connect to the RPA and attach it to the
cluster. Attaching the RPA to the cluster, reboots the RPA. It takes a couple of minutes before you can connect to the RPA again. The status information in the GUI indicates that the RPA is up (status changes from Error to Ok). Attach RPA to Cluster
End of Lab Exercise
** Cluster operations ** [1] Attach to cluster [2] Detach from cluster [B] Back [Q] Quit pod1 RPA1: 1 Do you want to attach the RPA to the cluster (Note: RPA will be rebooted) (y/n)? y You have configured an NTP server for the RPA cluster. There is no need to set the system clock. Do you want to do so anyway? (y/n)? n RPA is going to be attached and reboot. Attach to cluster succeeded
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 213 of 223
Part 4: Running Log Collection Step Action
1 Log into your RPA as boxmgmt. From the Main menu, select Diagnostics then Collect System Info. The RPA displays the current GMT. Always perform log collection in GMT regardless of your location. Select a start date and time, then select an end date and time. You may choose to collect logs from the remote site. Note: When collecting logs for support it is best to gather as much information as possible.
Collect System Info
** Diagnostics ** [1] IP diagnostics [2] Fibre Channel diagnostics [3] Synchronization diagnostics [4] Collect system information [B] Back [Q] Quit pod1 RPA1: 4 GMT right now is: 09/26/2009 00:16:57 Enter the date from which to start collecting information (MM/dd/yyyy) (Default '09/25/2009'): Enter the time from which to start collecting information (HH:mm:ss) (Default '20:16:57'): Enter the date at which to stop collecting information (MM/dd/yyyy) (Default '09/26/2009'): Enter the time at which to stop collecting information (HH:mm:ss) (D f lt '00 16 57')
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 214 of 223
Step Action 2 In the lab environment, enter n when prompted to send the files to an ftp server. You may
choose to collect logs from the RPAs, hosts, and SANTap. The more components selected the longer the log collection may take. If you choose to limit the log collection time, some information may not be gathered. If you selected to collect logs from hosts, you are prompted to choose which hosts. This allows you to focus the collection on just the system(s) involved in the issue. Collect System Info – Collection Mode
3 Once the log collection is completed, the results are printed to the screen. Notice the IP address for accessing the info page where the zipped log files are stored.
Collect System Info - Completion
Do you want to copy the output file to a remote FTP server (y/n)? n Do you want to include core files in the collection process (y/n)? n Do you want to include both sites in the collection process (y/n)? y 1. Collect system information from RPAs only 2. Collect system information from splitters only 3. Collect system information from RPAs and splitters Select the system information that you want to include in the collection process (1 to 3): 3 Do you want to collect system information from all of the splitters (Default is 'y') (y/n)? n ... To collect system information from RPAs that returned an error, try opening a direct SSH connection to each RPA that returned an error, and initiating the Collect System Information process from the Installation Manager of that RPA. The specified system information has been successfully collected. Browse to http://10.127.14.22/info/ or https://10.127.14.22/info/ and log in as a user with webdownload privileges to retrieve the output file.
** Diagnostics ** [1] IP diagnostics [2] Fibre Channel diagnostics [3] Synchronization diagnostics [4] Collect system information [B] Back [Q] Quit pod1 RPA1: 4 GMT right now is: 09/26/2009 00:16:57 Enter the date from which to start collecting information (MM/dd/yyyy) (Default '09/25/2009'): Enter the time from which to start collecting information (HH:mm:ss) (Default '20:16:57'): Enter the date at which to stop collecting information (MM/dd/yyyy) (Default '09/26/2009'): Enter the time at which to stop collecting information (HH:mm:ss) (Default '00:16:57'):
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 215 of 223
Step Action 4 Open a web browser and enter the IP address of the location provided by the log collection
output. When prompted, login as user webdownload with the password webdownload.
Accessing Collected Log Files – Webdownload Login
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 216 of 223
Step Action 5 After logging in, you see the Index of /info. There may be multiple log collections to
choose from. Double click on the file you want and choose to open or save.
Index of /info
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 217 of 223
Step Action 6 Scroll through the archive to display the files collected. A log collection can be very large
and include a considerable number of files.
Viewing Zipped Logs
End of Lab Exercise
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 218 of 223
Part 5: Performance and Statistics Step Action
1 Connect to the site management IP address as admin. Example: ssh [email protected] Note: If your lab setup contains only one RPA, the management IP address will not be applicable. Connect using the box management IP.
2 Run the detect_bottlenecks command. Select the Mode, the earliest time and latest time available when prompted.
Running detect_bottlenecks – Selecting Mode and Time Range
pod2> detect_bottlenecks Enter mode (select one from the list), or press Enter to default mode (default is System overview and bottleneck analysis) 1) System overview and bottleneck analysis 2) Detection of initialization periods 3) Detection of high load periods 4) General detection including initialization and high load periods with peak writing analysis 5) Advanced and detailed general detection 6) Peak writing analysis Select, or press 'ENTER': 1 Enter earliest time, or press 'ENTER' for the earliest time available (Format: HH:MM [DD/MM/YYYY] OR wks/days/hrs/mins ago)
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 219 of 223
Step Action 3 Select the type of overview, the time duration for I/O peaks, and the consistency group
names. Enter yes to confirm when prompted.
Running detect_bottlenecks – Refining Results
4 View the results.
Command Output Example – detect_bottlenecks
Statistics were found between the times: 2009/09/07 15:56:26.535 GMT ----> 2009/09/26 00:34:04.314 GMT ================================================================================ System overview of site: pod1 WAN throughput from site : 0.00237323 Megabits/sec Max Value: 0.00349196 Megabits/sec Total incoming writes rate for site : 1.57604e-07 Megabytes/sec Max Value: 0.000386758 Megabytes/sec Incoming IOs rate for site : 0.82074 IOs /sec Max Value: 3.33699 IOs /sec Initialization output rate for site (average over all period) : 0 Megabytes /sec --More--(42%)Bottlenecks detection file successfully written to the disk.
Do you want advanced overview? (default is NO) 1) yes 2) no Select, or press 'ENTER': Do you want detailed overview? (default is NO) 1) yes 2) no Select, or press 'ENTER': Enter peak_duration(secs/mins/hrs),e.g I/O peaks (default is 5mins) Enter group name(s), separated by ',' if more than one (default is all groups) Warning: This operation may take a long time. Do you want to continue (yes/no)? [default is 'yes']
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 220 of 223
Step Action 5 From the admin command prompt, enter get_system_statistics. Type in the site name or
press enter for both sites.
Running get_system_statistics
6 From the admin command prompt, enter get_group_statistics. Enter the name of your consistency group or press Enter for all groups.
Running get_group_statistics
pod2> get_group_statistics Enter consistency group name, or press 'ENTER' for all groups Group: CG_Lab: Copy stats: Nw_Copy: Journal: Usage: 238.57MB Total: 2.45GB Latest image: pre-replication image Current image: pre-replication image Journal lag: 0.00B Protection window:N/A Average compression:N/A Mode: Normal PC_Lab: SAN traffic: Data: 0 bps
Writes (per sec): 0
pod2> get_system_statistics Enter site name, or press 'ENTER' for both sites Do you want summarized site statistics (as opposed to full box statistics)? (default is no) 1) yes 2) no Select, or press 'ENTER': 1 RPA statistics: None Site statistics: pod1: Application: SAN: 0 bps WAN: 1904 bps Application (writes): 0 Compression: 0 pod2: Application: SAN: 0 bps WAN: 1848 bps Application (writes): 0
Compression: 0
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 221 of 223
Step Action 7 From the admin command prompt, enter export_statistics. Select the earliest and latest
times and the desired criteria (or enter for the default values). Enter a file name when prompted and confirm execution by entering y.
Running export_statistics
pod2> export_statistics Enter earliest time, or press 'ENTER' for no time filter (Format: HH:MM [DD/MM/YYYY] OR wks/days/hrs/mins ago) Enter latest time, or press 'ENTER' for now (Format: HH:MM [DD/MM/YYYY] OR wks/days/hrs/mins ago) Include global statistics? (default is yes) 1) yes 2) no Select, or press 'ENTER': 1 Enter site name, or press 'ENTER' for both sites Enter RPA name or a list of RPAs separated by ','. Enter group name(s), separated by ',' if more than one (default is all groups. print 'none' to get global only) Enter categories filter (available categories are: compression, highload, init, overview, performance, regulation, wan) (default is overview) Enter frequency (mins/hrs) with which to extract statistics. Enter file name to which you want to export statistics (or press ENTER for default file. Legal file name may contain letters, digits, '_', '.' and '-') Warning: This can take a long time. Do you want to continue(y/n)?y
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 222 of 223
Step Action 8 When the export is completed, open a web browser and access the info page for the RPA as
user webdownload.
Accessing /info
9 On the Index of /info, click the folder name you designated during the export.
Index of /info
RecoverPoint Implementation Lab Guide
Copyright © 2009 EMC Corporation. All Rights Reserved. Version 2.2 Page 223 of 223
Step Action 10 A list of exported files are displayed. Click one of the files to open in Excel and view the
contents.
Viewing Index of Files
End of Lab Exercise