Currently a backup operation can only be performed when the volume is detached. This blueprint proposes the following changes in order to do a non-disruptive backup while the volume is still attached: For the LVM driver, a temporary snapshot will be created and backup will be performed by reading from the local path of the snapshot. After that, the temporary snapshot will be deleted. For other drivers, a default implementation will be provided to create a temporary volume from the source volume and backup will be performed by attaching the temporary volume and reading from it. After it is done, the temporary volume will be detached and deleted. For other drivers, a more optimized approach will also be provided to create a temporary snapshot, attach the snapshot, backup the snapshot, and then detach and delete the temporary snapshot. This optimized approach is not required. If a driver has not implemented attach snapshot, the default implementation with a temporary volume will be used. DocImpact blueprint non-disruptive-backup Change-Id: I069d86d315dd2c83d43d694087f2dcc2f0b0752b
7.6 KiB
Non Disruptive Backup
https://blueprints.launchpad.net/cinder/+spec/non-disruptive-backup
Currently a backup operation can only be performed when a volume is detached. This blueprint proposes to use a temporary snapshot or volume to do non-disruptive backup.
Problem description
If a volume is attached, it has to be detached first before it can be backed up. This is not efficient and it disrupts normal operations by cloud users just to do a backup.
Use Cases
The use case is that an end user would like to do a backup without detaching the volume.
Proposed change
For an attached volume, taking a temporary snapshot is usually less expensive than creating a whole temporary volume. If we can attach the snapshot and read it directly, we do not have to create another temporary volume. It is quicker to complete the backup this way.
If a driver has not implemented attach snapshot and does not have a way to read from a snapshot, we can create a temporary volume from the attached source volume, and backup the temporary volume.
For the LVM driver, because the backup service is on the same node as the volume driver, we can just read the local path of the temporary snapshot after it is created from the attached source volume. We will create a temporary snapshot, but there is no need to attach snapshot for LVM.
In the future, if the backup service can be on a different node than the volume driver, then we can look into whether to do attach snapshot for LVM as well, however, there are concerns that introducing the attach snapshot logic will complicate the LVM code. This is for the future, so it can be decided later.
This proposal will be implemented in multiple steps.
Step 1:
For the LVM driver, take a temporary snapshot of the attached volume and back it up.
- Take a temporary snapshot
- Do backup from the local path of the temporary snapshot
- Cleanup temporary snapshot
For other drivers, provide a default implementation by creating a temporary volume and backing it up.
- Create a temporary volume from the source volume
- Attach the temporary volume
- Do backup from the temporary volume
- Detach the temporary volume
- Cleanup temporary volume
Note: If the volume is not attached, it will be backed up the same way as before.
Step 2:
Provide a more optimized way for other drivers to use a temporary snapshot to backup an attached volume. For drivers that cannot read from a local path of a snapshot, we need to define an interface for attaching a snapshot, similar to attaching a volume.
- Take a temporary snapshot
- Attach the temporary snapshot
- Do backup from the temporary snapshot
- Detach the temporary snapshot
- Cleanup temporary snapshot
Step 3 (future):
If the backup service can use the volume driver on a remote node in the future, we can provide a way for the call to happen through RPC. This is similar to step 2, except that it will call the volume manager on a remote node using RPC API to attach snapshot.
Alternatives
There are a couple of alternatives: * Detach the volume and back it up. * Take a snapshot of the attached volume, create a volume from the snapshot and then back it up.
Data model impact
Step 1:
Add the following new columns to the backups table for the temporary volume and snapshot id:
temp_volume_id = Column(String(36)) temp_snapshot_id = Column(String(36))
Add the following new column to the volumes table to persist the status of a volume before it is changed to 'backing-up' status. This will be used to restore the volume status back to either 'available' or 'in-use' after the backup is complete. This is also used for cleaning up failed backups when the backup service is restarted. This is because the cleanup procedure is different for 'available' and 'in-use' volumes.
previous_status = Column(String(255))
Step 2:
Add the following new column to the snapshots table. The snapshots table already has provider_id and provider_location like the volumes table. This provider_auth column is needed to save the authentication data when a target is created for the snapshot. This is needed if we want to attach the snapshot.
provider_auth = Column(String(255))
REST API impact
Step 1:
Change the existing create backup API to take a force flag. If the volume is 'in-use', the force flag has to be True. By default it is False. The force flag is not needed for 'available' volumes.
- Create backup
V2/<tenant id>/backups
Method: POST
JSON schema definition for V2:
{ "backup": { "display_name": "nightly001", # existing "display_description": "Nightly backup", # existing "volume_id": "xxxxxxxx", # existing "container": "nightlybackups", "force": True, # new } }
Step 2:
The following driver APIs will be added to support attach snapshot and detach snapshot.
attach snapshot: * def _attach_snapshot(self, context, snapshot, properties, remote=False) * def create_export_snapshot(self, conext, snapshot) * def initialize_connection_snapshot(self, snapshot, properties, ** kwargs)
detach snapshot: * def _detach_snapshot(self, context, attach_info, snapshot, properties, force=False, remote=False) * def terminate_connection_snapshot(self, snapshot, properties, ** kwargs) * def remove_export_snapshot(self, context, snapshot)
Alternatively we can use an is_snapshot flag for volume and snapshot to share common code without adding new functions, but it will make the code confusing and hard to read. So there is a trade off between reducing code duplication and increasing code readibilty here.
Security impact
None
Notifications impact
None
Other end user impact
End user will be able to create a backup without detaching the volume.
Performance Impact
No obvious performance impact.
If we can attach the snapshot and back it up with the proposed change, it will be cleaner and easier than manually taking a snapshot, creating a volume from the snapshot, and then backing it up and deleting it.
Other deployer impact
The deployer will be able to backup an attached volume.
Developer impact
Driver developers can implement the proposed new driver APIs for more efficient backup. This is not required though. A default implementation will be provided to create a temporary volume from the source volume as discussed earlier.
Implementation
Assignee(s)
- Primary assignee:
-
<xing-yang>
- Other contributors:
-
<None>
Work Items
- Step 1: Provide a default implementation. Patch in review here: https://review.openstack.org/#/c/193937/
- Step 2: Provide a more optimal implementation by adding new driver APIs to support attach snapshot. WIP patch proposed here: https://review.openstack.org/#/c/201249/
- Step 3 (future): This will happen after the backup service is decoupled from the volume driver.
Dependencies
None
Testing
Unit tests will be provided.
Documentation Impact
Documentation will be modified to describe this feature.
References
- Link to summit discussion: https://etherpad.openstack.org/p/cinder-liberty-backup-using-snapshot