High availability and disaster recovery
This chapter provides information about the opportunities for high availability (HA) when using VMware vSphere 4.1 and N series storage in the same environment. It then explains the implementation of disaster recovery using the functions of these technologies.
This chapter includes the following topics:
12.1 High availability
This section provides details about some of the high availability features of the N series and Virtual Infrastructure 3 solution.
12.1.1 N series node failures
In a normal configuration, two N series servers are clustered. If a failure occurs in one of the nodes, the second system automatically takes on the load of both servers without any manual intervention required.
However, if a failure affects both nodes, such as a power failure for the whole server environment, a disaster recovery implementation is required. This implementation can be in the form of a second pair of N series servers in a location nearby, using MetroCluster. Or it can be done with a pair of N series servers in a more remote location, using SnapMirror.
An N series cluster (standard N series configuration) offers the following high availability features:
Built-in redundancy for a failure of a power supply, fan, or disk controller
RAID-DP for a single or dual disk failure
Multipath for a single disk path or port failure
Snapshot copies for accidental erasure or destruction of data
MetroCluster is an extended N series cluster for distances of up to 100 km with fiber connectivity between sites. It provides the following additional HA features:
SyncMirror for a triple disk failure or complete disk shelf failure
Redundancy for a host bus adapter (HBA) or port failure
Active-active controller configuration for a storage controller failure
MetroCluster for a data center power or environmental outage
The ability of VMware HA cluster to be split across the MetroCluster
Figure 12-1 shows a fabric attached MetroCluster configuration.
Figure 12-1 MetroCluster configurations
12.1.2 VMware host failures
With two or more VMware hosts configured in a cluster with a shared storage, you can have high availability features. Virtual machines on a failed host can be quickly restarted on another host, as long as there is capacity available on the remaining hosts. This feature is enabled by VMware High Availability (HA). As a preferred practice, provide enough capacity on your environment for the failure of at least one host, also known as N+1. Depending on your availability requirements and the speed of growth of your environment, you might even want to size it N+2.
Another feature available is Dynamic Resource Scheduler (DRS), which manages the load of the guests across the servers in the cluster. If one of the hosts becomes overloaded, guests can be automatically moved to a server with a less load without any downtime. If you plan to use the VMware HA feature, you can also use the DRS feature. This feature allows virtual machines to be evenly balanced across the cluster in the event of a host failure.
If you do not have high availability on your environment, use operating system or application-level clustering. If your application is not state-aware, use load balancers, as for web servers.
12.2 Disaster recovery options
You can mirror an N series node (cluster) at the primary site to an N series node at a secondary site (Figure 12-2). It can be used in a development or test capacity during normal operation if the loss of it in a disaster is acceptable. Otherwise, it can be used for on demand or out-of-band additional capacity.
Disaster recovery can also be done using a FlexClone of the SnapMirror. You can even start the virtual machines in the DR site while the run on the primary site if their network is isolated. This method uses a lot less disk than traditional methods, because cloning does not require a full copy of the source, but rather only as changes occur on either copy.
A VMware host or cluster must be in the disaster recover site also to run the VMs present on the cloned storage at DR site. However, it does not have to be the same hardware, thus providing more flexibility to your planning.
Figure 12-2 N series Gateway cluster configuration
12.3 Setting up disaster recovery
In this section, you configure a Virtual Infrastructure 3 and N series environment to use the N series SnapMirror feature. This feature provides replication of the datastores to a second location that is ready for use in the event of a disaster.
The following tasks are involved:
1. Configuring the source location storage
2. Enabling SnapMirror on the N series storage systems
3. Configuring the mirror
4. Starting the mirror
The SnapMirror configuration is similar in many ways to SnapVault configuration. Therefore, if you already reviewed Chapter 11, “Backup and recovery to a separate system” on page 197, you can see that the setup is familiar.
12.3.1 Setting up the primary storage
If you are setting up a new environment, you can plan your storage based on your disaster recovery requirements. Where possible, co-locate data with similar disaster recovery requirements on the same volumes. More importantly, try not to store data with separate requirements on the same volume. For example, make sure that your transient data is stored on separate volumes from your vital data.
To set up the primary storage, follow these steps:
1. Set up your primary storage as for any N series storage for VMware.
2. On the destination storage system, create a volume for each volume you intend to replicate that is at least as large as the source volume. However, do not create LUNs, because they are replicated from the source.
3. Restrict access to the destination volumes by entering the vol restrict <vol_name> command (Example 12-1). This command prevents the volume from being accessed by the virtual machines outside of a disaster situation.
Example 12-1 Restricting a destination volume
itsotuc1> vol restrict vol_vm_dr
Volume 'vol_vm_dr' is now restricted.
itsotuc1>
4. On the destination storage system, create a volume with the appropriate LUNs that are the same as each of the volumes on the source that contains the transient data.
5. Disable the automatic snapshots of both the source and destination volumes unless you have a separate need for them.
 
SnapMirror: Unlike SnapVault, which requires Qtrees, SnapMirror works at either the Qtree level or volume level. The examples in this section use volumes, but you can use Qtrees instead if you prefer.
12.3.2 Licensing SnapMirror
To use SnapMirror, you must apply your site license to the source and destination N series storage systems and to the clustered nodes for each system, if applicable:
1. In FilerView, in the left navigation pane, select Filer → Manage Licenses.
2. In the Manage Licenses pane (Figure 12-3), enter your license code and select Apply.
Figure 12-3 SnapMirror License installed
When installed, the SnapMirror options become available in the left navigation pane (Figure 12-4) of FilerView.
Figure 12-4 SnapMirror menu options
12.3.3 Setting permissions
Set the permissions to allow the destination system to access SnapMirror on the source by entering the following command on the source system (Example 12-2):
options snapmirror.access host=<secondary>
Example 12-2 Setting the SnapVault permissions
itsotuc4*> options snapmirror.access host=9.11.218.114
itsotuc4*> options snapmirror.access
snapmirror.access host=9.11.218.114
itsotuc4*>
The options snapmirror.access command verifies that the permission was assigned correctly.
You can also use this function in FilerView. In the left navigation pane, select SnapMirror → Remote Access → Add. However, use the CLI command shown in Example 12-2 to confirm that the access was assigned correctly.
12.3.4 Configuring the volume mirror
To configure the volume mirror, follow these steps:
1. Set up the mirror transfer from the secondary system. In FilerView, in the left navigation pane (Figure 12-5), select SnapMirror → Add.
Figure 12-5 Selecting the option to add SnapMirror
2. In the Destination Location panel of the SnapMirror Wizard (Figure 12-6), select the destination volume you created for this volume mirror. Then click Next.
Figure 12-6 SnapMirror destination
3. In the Source Location panel shown in Figure 12-7, enter the IP address (or DNS name if you prefer) of the source N series system, and the volume you want to mirror. Then click Next.
Figure 12-7 IP address of the remote storage
4. In the Set Schedule panel (Figure 12-8), limit the transfer rate by selecting how often you want transfer updates to occur, based on your disaster recovery requirements. If you do not want to set any limits, select Never. Then click Next.
Figure 12-8 SnapMirror schedule
5. In the Commit panel (Figure 12-9), verify that the settings you entered are correct. Then click Commit.
Figure 12-9 SnapMirror implementation summary
6. Verify that the SnapMirror was added successfully (Figure 12-10).
Figure 12-10 SnapMirror added successfully
12.3.5 Starting a mirror
After you configure the mirror, you must initialize it to start the initial mirror copy to the destination storage system:
1. In FilerView on the destination system, in the left navigation pane (Figure 12-11), select SnapMirror → Report.
2. In the SnapMirror Report pane (Figure 12-11), select the SnapMirror that you configured. Notice that it is currently not initialized.
Figure 12-11 SnapMirror not initialized
3. In the SnapMirror Properties panel (Figure 12-12), click Initialize to start the initial SnapMirror between the two storage systems.
Figure 12-12 Initializing SnapMirror
4. In the SnapMirror Properties pane, verify that the initialization process started successfully.
5. Check the SnapMirror Report (Figure 12-13) again for the status. The SnapMirror is idle, because the mirror has been created. Also no scheduled processes are running. You see a similar report on the source server.
Figure 12-13 Checking the SnapMirror Report
You can also check the SnapMirror status in the Manage Snapshots menu:
a. In the left navigation pane of FilerView (Figure 12-14), select Volumes → SnapShots → Manage.
Figure 12-14 Selecting to manage snapshots
b. In the Manage Snapshots pane (Figure 12-15), select the SnapMirror volume, and click View to see the snapshots for that volume.
Figure 12-15 SnapMirror in FilerView
12.4 Recovering from a disaster
If a disaster (or possibly a full test of the disaster recovery capability) occurs, perform the following tasks:
1. Break the mirror to make the mirrored data writable.
2. Map the LUNs.
3. Rescan the VMware hosts to see the LUNs.
4. Reinventory the virtual machines.
5. Start the virtual machines.
12.4.1 Breaking the mirror
During the setup procedure, the mirror volumes in the destination location were restricted to prevent writes. To remove this restriction and allow the data to be mounted and accessed, break the mirror:
1. Run FilerView on the destination N series system.
2. In the left navigation pane of FilerView (Figure 12-16), select SnapMirror → Report.
Figure 12-16 SnapMirror Report
3. In the SnapMirror Report pane (Figure 12-17), select the volume you want to use.
Figure 12-17 Selecting the volume
4. In the SnapMirror Properties pane (Figure 12-18), where you see the properties of this volume replica, click Quiesce to ensure that no data is unwritten.
Figure 12-18 Quiescing the volume
5. When the quiesce is successful, click Break (Figure 12-19).
Figure 12-19 Breaking the mirror
6. Verify that the break operation completed successfully.
7. Repeat these steps for each mirrored volume that you require access to on the destination system.
12.4.2 Mapping the LUNs and rescanning VMware hosts
Now that the mirror is broken and the data is available, any LUNs on the volume must be mapped so that the VMware host can use them.
1. Map the LUN as already previously explained.
2. Create a datastore using the LUN you just mapped.
3. Then reinventory the virtual machines.
Also see the script provided in Appendix A, “Hot backup Snapshot script” on page 279 to help you to perform the tasks.
12.4.3 Starting virtual machines
Now that the virtual machines are configured correctly, start them:
1. Right-click a virtual machine and select Power, then Power On
2. On the right side of the window, when prompted for the Unique Identifier (UUID) settings, select Create (Figure 12-20), and click OK.
Figure 12-20 Creating a UUID
3. Verify the task list to confirm that the guest started correctly (Figure 12-21).
Figure 12-21 Guest started successfully
4. Repeat these steps for each guest you want to start in the DR environment. You might also want to start the remote console for the guests, or run application diagnostic tests for each application, to confirm that everything is working as expected.
12.5 Returning to production
In a case where a disaster occurred and the environment is failed over to the disaster recovery site, the data stored in there is the most current. If the production environment comes back online later, the data and server load might need to be transferred back. Similar to regular SnapMirror transfers, the production site can be updated from the disaster recovery data while the disaster recovery site is operational. This update might be large if the production data was lost or corrupted, or it might be small if the production data was unaffected by the disaster. The server load change requires an outage. Therefore, it is better to schedule this outage to occur in non-productions hours.
Returning to production entails the following high-level procedure:
1. Repair or recover the production N series storage system to its original state, with the correct software, options, and so on.
2. Copy the data (or changes) back to the production site from the disaster recovery site while the disaster recovery system is operational for users.
3. Prevent users or systems from accessing the disaster recovery data, and copy any final updates to production.
4. Split the mirror between the two sites.
5. Remap the production LUNs.
6. Rescan the VMware hosts, and inventory the virtual machines.
7. Start the virtual machines.
8. Re-establish SnapMirror from production to the disaster recovery site.
Because many of these steps are the same as in the disaster scenario, only the new steps are explained in detail in this section.
 
FilerView versus CLI commands: It is possible to perform some of the steps in this section and the following sections from FilerView. However, some are not available as they are not commonly performed operations. As a result, they are all shown as CLI commands.
12.5.1 Replicating data from disaster recovery to the production site
After the production site N series server becomes available, copy the data from the disaster recovery N series system to the production system. You can do this task by using one of the procedures in the following sections, depending on the state of the production N series data.
Before you begin, assign permissions in the reverse direction of what is explained in 12.3.3, “Setting permissions” on page 220, but enter the following command:
options snapmirror.access host=<secondary>
Production N series data still intact
If the data in the production site was not lost, you need only to copy updates back from the disaster recovery site. You can perform this task by entering the following command:
snapmirror resync -S <DR_syste,m>:<volume> <prod_system>:<volume>
Example 12-3 shows the execution of the snapmirror command.
Example 12-3 Synchronizing the production N series with disaster recovery updates
itsotuc4>snapmirror resync -S 9.11.218.114:vol_vm_dest itsotuc4:vol_vm_source
The resync base Snapshot will be: itsotuc1(0118052508)_vol_vm_dest.5
Are you sure you want to resync the volume? yes
Thu May 1 23:30:55 MST last message repeated 2 times
Thu May 1 23:30:58 MST [itsotuc4: snapmirror.dst.resync.info:notice]: SnapMirror resync of vol_vm_source to 9.11.218.114:vol_vm_dest is using itsotuc1(0118052508)_vol_vm_dest.5 as the base Snapshot.
Volume vol_vm_source will be briefly unavailable before coming back online.
Thu May 1 23:30:59 MST [itsotuc4: wafl.snaprestore.revert:notice]: Reverting volume vol_vm_source to a previous Snapshot.
Thu May 1 23:30:59 MST [itsotuc4: wafl.vol.guarantee.replica:info]: Space for replica volume 'vol_vm_source' is not guaranteed.
Revert to resync base Snapshot was successful.
Thu May 1 23:30:59 MST [itsotuc4: snapmirror.dst.resync.success:notice]: SnapMirror resync of vol_vm_source to 9.11.218.114:vol_vm_dest successful.
Transfer started.
Monitor progress with 'snapmirror status' or the snapmirror log.
itsotuc4>
Production N series recovery
If the data in the production site was lost or corrupted during the disaster situation, you must re-create the volumes and then copy back all of the data from the disaster recovery site. You re-create the volume in the production site, and restrict the volume. Initialize the production system from the good copy on the disaster recovery system by entering the following command on the production N series system:
snapmirror initialize -S <dr_system>:<dr_vol> <prod_system>:<prod_vol>
Example 12-4 shows the snapmirror initialize command.
Example 12-4 Copying the disaster recovery environment data to the production site
itsotuc4> snapmirror initialize -S 9.11.218.114:vol_vm_dr itsotuc4:vol_vm_dr
Transfer started.
Monitor progress with 'snapmirror status' or the snapmirror log.
After the initialization is complete, the production system has a copy of the data again.
12.5.2 Preventing access and performing a final update
To ensure that the data is up to date, all virtual machines running on the disaster recovery site N series system must be shut down. Shutting down this system ensures that the final updates of data can be transferred back to the production system.
If a time lag exists between when the initialization was started and when it is convenient to schedule an outage on the guests, perform an update while the virtual machines are still running. Then shut down all guests that are accessing the disaster recovery site data.
When there is no longer anything accessing the DR site data, run the following command from the production N series system to perform the update:
snapmirror update -S <dr_system>:<dr_vol> <prod_system>:<prod_vol>
Example 12-5 shows the results of the snapmirror update command.
Example 12-5 Updating data between the disaster recovery and production sites
itsotuc4> snapmirror update -S 9.11.218.114:vol_vm_dr itsotuc4:vol_vm_dr
Transfer started.
Monitor progress with 'snapmirror status' or the snapmirror log.
itsotuc4>
12.5.3 Splitting the mirror
Now both the disaster recovery and production systems have the same data, and no changes are occurring on either system. Therefore, the mirror can be broken.
From the production N series system, quiesce and break the mirror by using the following command:
snapmirror break <volume_name>
Example 12-6 shows the execution of the snapmirror break command.
Example 12-6 Breaking the mirror
itsotuc4> snapmirror break vol_vm_dr
snapmirror break: Destination vol_vm_dr is now writable.
Volume size is being retained for potential snapmirror resync. If you would like to grow the volume and do not expect to resync, set vol option fs_size_fixed to off.
itsotuc4>
12.5.4 Re-establishing the mirror from the production to disaster recovery site
Finally, you can perform a resynchronization to make the disaster recovery site a mirror of the production site again. Enter the following command on the disaster recovery N series system:
snapmirror resync <vol_name>
Example 12-7 shows the results of the snapmirror resync command.
Example 12-7 Resync from the production to disaster recovery site
itsotuc1> snapmirror resync vol_vm_dr
The resync base Snapshot will be: itsotuc4(0101165597)_vol_vm_dr.2
Are you sure you want to resync the volume? yes
Thu May 1 16:32:15 MST [snapmirror.dst.resync.info:notice]: SnapMirror resync of vol_vm_dr to 9.11.218.238:vol_vm_dr is using itsotuc4(0101165597)_vol_vm_dr.2 as the base Snapshot.
Volume vol_vm_dr will be briefly unavailable before coming back online.
Thu May 1 16:32:16 MST [wafl.snaprestore.revert:notice]: Reverting volume vol_vm_dr to a previous Snapshot.
Thu May 1 16:32:16 MST [wafl.vol.guarantee.replica:info]: Space for replica volume 'vol_vm_dr' is not guaranteed.
Revert to resync base Snapshot was successful.
Thu May 1 16:32:16 MST [snapmirror.dst.resync.success:notice]: SnapMirror resync of vol_vm_dr to 9.11.218.238:vol_vm_dr successful.
Transfer started.
Monitor progress with 'snapmirror status' or the snapmirror log.
itsotuc1>
12.5.5 Configuring VMware hosts and virtual machines on the production site
Now the production N series system is the source again, and replication is occurring back to the disaster recovery site. Perform the following steps to start the guests on the production VMware hosts:
1. Rescan the VMware hosts to view the datastores again.
The new datastore might be displayed as a snapshot. Therefore, you can rename it to the original name before using it, as in Figure 12-22.
Figure 12-22 Recovered datastore
2. Reinventory the virtual machines.
You might need to delete the original virtual machines first.
3. Reconfigure the virtual machines for the transient data volumes of the production site.
4. Start the virtual machines.
12.6 Disaster recovery testing
In a disaster recovery test, it is often desirable to perform testing without disrupting either the source environment or the destination copy of the data. Such a test is relatively easy to perform with the use of N series cloning, so that the disaster recovery environment can be tested against a clone of the mirrored data. Similar to other N series cloning processes, the clone requires little additional disk capacity in the disaster recovery site, because only changes are written to disk.
To perform this type of test, the LAN environment for the disaster recovery VMware hosts must be separated from the production environment. Thus, the guests can be started without causing conflicts in the network. You can complete this task by isolating the VMware hosts from the network (while still providing connectivity to the N series server). Alternatively, if feasible, you can set up isolated virtual networks within the VMware hosts. This second option, however, prevents communication between guests on separate hosts.
You can perform a disaster recovery test with N series cloning by using the following high-level procedure:
1. Verify that SnapMirror Snapshots in the disaster recovery location are current.
2. Clone the Snapshot volumes.
3. Bring the cloned LUNs online, and map them for access by the VMware hosts.
4. Rescan the VMware hosts.
5. Add the virtual machines to the inventory.
6. Start the virtual machines.
7. Perform disaster recovery application testing.
8. When complete, stop the virtual machines, remove them from the inventory, and destroy the cloned volumes.
 
..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset
3.133.160.156