IBM XIV/DS8K: Implement Replication v2.1

This patch implements cinder Replication version 2.1 as proposed in
https://blueprints.launchpad.net/cinder/+spec/replication

The driver implements:
1. freeze_backend - notify the backend that it's frozen
2. thaw_backend - notify the backend that it's not frozen
3. host_failover - failover a backend to a secondary replication target

Also removed v1.0 and v2.0 functions

DocImpact
Change-Id: I50498180c63725ab0f524ff8d55952446800570f
Closes-Bug: #1543600
This commit is contained in:
Alon Marx 2016-03-03 18:15:38 +02:00
parent 11be817356
commit 9952531da4
3 changed files with 73 additions and 268 deletions

View File

@ -85,7 +85,8 @@ CONNECTOR = {'initiator': "iqn.2012-07.org.fake:01:948f189c4695", }
class XIVDS8KFakeProxyDriver(object): class XIVDS8KFakeProxyDriver(object):
"""Fake IBM XIV and DS8K Proxy Driver.""" """Fake IBM XIV and DS8K Proxy Driver."""
def __init__(self, xiv_ds8k_info, logger, expt, driver=None): def __init__(self, xiv_ds8k_info, logger, expt,
driver=None, active_backend_id=None):
"""Initialize Proxy.""" """Initialize Proxy."""
self.xiv_ds8k_info = xiv_ds8k_info self.xiv_ds8k_info = xiv_ds8k_info
@ -169,32 +170,11 @@ class XIVDS8KFakeProxyDriver(object):
return (self.volumes[volume['name']].get('attached', None) return (self.volumes[volume['name']].get('attached', None)
== connector) == connector)
def reenable_replication(self, context, volume):
model_update = {}
if volume['replication_status'] == 'inactive':
model_update['replication_status'] = 'active'
elif volume['replication_status'] == 'invalid_status_val':
raise exception.CinderException()
model_update['replication_extended_status'] = 'some_status'
model_update['replication_driver_data'] = 'some_data'
return model_update
def get_replication_status(self, context, volume): def get_replication_status(self, context, volume):
if volume['replication_status'] == 'invalid_status_val': if volume['replication_status'] == 'invalid_status_val':
raise exception.CinderException() raise exception.CinderException()
return {'replication_status': 'active'} return {'replication_status': 'active'}
def promote_replica(self, context, volume):
if volume['replication_status'] == 'invalid_status_val':
raise exception.CinderException()
return {'replication_status': 'inactive'}
def create_replica_test_volume(self, volume, src_vref):
if volume['size'] != src_vref['size']:
raise exception.InvalidVolume(
reason="Target and source volumes have different size.")
return
def retype(self, ctxt, volume, new_type, diff, host): def retype(self, ctxt, volume, new_type, diff, host):
volume['easytier'] = new_type['extra_specs']['easytier'] volume['easytier'] = new_type['extra_specs']['easytier']
return True, volume return True, volume
@ -284,42 +264,24 @@ class XIVDS8KFakeProxyDriver(object):
return {'status': 'deleted'}, updated_snapshots return {'status': 'deleted'}, updated_snapshots
def replication_disable(self, context, volume): def freeze_backend(self, context):
if volume['replication_status'] == 'invalid_status_val': return True
raise exception.CinderException()
volume_model_update = {} def thaw_backend(self, context):
volume_model_update['replication_driver_data'] = FAKE_DRIVER_DATA return True
return volume_model_update def failover_host(self, context, volumes, secondary_id):
target_id = 'BLA'
volume_update_list = []
for volume in volumes:
status = 'failed-over'
if volume['replication_status'] == 'invalid_status_val':
status = 'error'
volume_update_list.append(
{'volume_id': volume['id'],
'updates': {'replication_status': status}})
def replication_enable(self, context, volume): return target_id, volume_update_list
if volume['replication_status'] == 'invalid_status_val':
raise exception.CinderException()
volume_model_update = {}
volume_model_update['replication_driver_data'] = FAKE_DRIVER_DATA
return volume_model_update
def list_replication_targets(self, context, volume):
targets = []
for target in REPLICATION_TARGETS:
targets.append({'type': 'managed',
'target_device_id': target['target_device_id']})
return {'volume_id': volume['id'], 'targets': targets}
def replication_failover(self, context, volume, secondary):
if volume['replication_status'] == 'invalid_status_val':
raise exception.CinderException()
model_update = {'host': FAKE_FAILOVER_HOST,
'provider_location': FAKE_PROVIDER_LOCATION,
'replication_driver_data': FAKE_DRIVER_DATA}
return model_update
class XIVDS8KVolumeDriverTest(test.TestCase): class XIVDS8KVolumeDriverTest(test.TestCase):
@ -512,40 +474,6 @@ class XIVDS8KVolumeDriverTest(test.TestCase):
VOLUME, VOLUME,
existing_ref) existing_ref)
def test_reenable_replication(self):
"""Test that reenable_replication returns successfully. """
self.driver.do_setup(None)
# assume the replicated volume is inactive
replicated_volume = copy.deepcopy(REPLICATED_VOLUME)
replicated_volume['replication_status'] = 'inactive'
model_update = self.driver.reenable_replication(
CONTEXT,
replicated_volume
)
self.assertEqual(
model_update['replication_status'],
'active'
)
self.assertTrue('replication_extended_status' in model_update)
self.assertTrue('replication_driver_data' in model_update)
def test_reenable_replication_fail_on_cinder_exception(self):
"""Test that reenable_replication fails on driver raising exception."""
self.driver.do_setup(None)
replicated_volume = copy.deepcopy(REPLICATED_VOLUME)
# on purpose - set invalid value to replication_status
# expect an exception.
replicated_volume['replication_status'] = 'invalid_status_val'
self.assertRaises(
exception.CinderException,
self.driver.reenable_replication,
CONTEXT,
replicated_volume
)
def test_get_replication_status(self): def test_get_replication_status(self):
"""Test that get_replication_status return successfully. """ """Test that get_replication_status return successfully. """
@ -579,66 +507,6 @@ class XIVDS8KVolumeDriverTest(test.TestCase):
replicated_volume replicated_volume
) )
def test_promote_replica(self):
"""Test that promote_replica returns successfully. """
self.driver.do_setup(None)
replicated_volume = copy.deepcopy(REPLICATED_VOLUME)
# assume the replication_status should be active
replicated_volume['replication_status'] = 'active'
model_update = self.driver.promote_replica(
CONTEXT,
replicated_volume
)
# after promoting, replication_status should be inactive
self.assertEqual(
model_update['replication_status'],
'inactive'
)
def test_promote_replica_fail_on_cinder_exception(self):
"""Test that promote_replica fails on CinderException. """
self.driver.do_setup(None)
replicated_volume = copy.deepcopy(REPLICATED_VOLUME)
# on purpose - set invalid value to replication_status
# expect an exception.
replicated_volume['replication_status'] = 'invalid_status_val'
self.assertRaises(
exception.CinderException,
self.driver.promote_replica,
CONTEXT,
replicated_volume
)
def test_create_replica_test_volume(self):
"""Test that create_replica_test_volume returns successfully."""
self.driver.do_setup(None)
tgt_volume = copy.deepcopy(VOLUME)
src_volume = copy.deepcopy(REPLICATED_VOLUME)
tgt_volume['size'] = src_volume['size']
model_update = self.driver.create_replica_test_volume(
tgt_volume,
src_volume
)
self.assertTrue(model_update is None)
def test_create_replica_test_volume_fail_on_diff_size(self):
"""Test that create_replica_test_volume fails on diff size."""
self.driver.do_setup(None)
tgt_volume = copy.deepcopy(VOLUME)
src_volume = copy.deepcopy(REPLICATED_VOLUME)
self.assertRaises(
exception.InvalidVolume,
self.driver.create_replica_test_volume,
tgt_volume,
src_volume
)
def test_retype(self): def test_retype(self):
"""Test that retype returns successfully.""" """Test that retype returns successfully."""
@ -985,113 +853,64 @@ class XIVDS8KVolumeDriverTest(test.TestCase):
volumes_model_update['status'], volumes_model_update['status'],
"volumes list status failed") "volumes list status failed")
def test_replication_disable(self): def test_freeze_backend(self):
"""Disable replication on the specified volume.""" """Test that freeze_backend returns successful"""
self.driver.do_setup(None) self.driver.do_setup(None)
ctxt = context.get_admin_context() # not much we can test here...
self.assertTrue(self.driver.freeze_backend(CONTEXT))
replicated_volume = copy.deepcopy(REPLICATED_VOLUME) def test_thaw_backend(self):
replicated_volume['replication_status'] = 'enabled' """Test that thaw_backend returns successful"""
volume_model_update = ( self.driver.do_setup(None)
self.driver.replication_disable(ctxt, replicated_volume))
self.assertTrue('replication_driver_data' in volume_model_update) # not much we can test here...
self.assertTrue(self.driver.thaw_backend(CONTEXT))
def test_replication_disable_fail_on_cinder_exception(self): def test_failover_host(self):
"""Test that replication_disable fails on driver raising exception.""" """Test that failover_host returns expected values"""
self.driver.do_setup(None) self.driver.do_setup(None)
replicated_volume = copy.deepcopy(REPLICATED_VOLUME) replicated_volume = copy.deepcopy(REPLICATED_VOLUME)
# on purpose - set invalid value to replication_status # assume the replication_status is active
# expect an exception.
replicated_volume['replication_status'] = 'invalid_status_val'
self.assertRaises(
exception.CinderException,
self.driver.replication_disable,
CONTEXT,
replicated_volume
)
def test_replication_enable(self):
"""Enable replication on the specified volume."""
self.driver.do_setup(None)
ctxt = context.get_admin_context()
replicated_volume = copy.deepcopy(REPLICATED_VOLUME)
replicated_volume['replication_status'] = 'disabled'
volume_model_update = (
self.driver.replication_enable(ctxt, replicated_volume))
self.assertTrue('replication_driver_data' in volume_model_update)
def test_replication_enable_fail_on_cinder_exception(self):
"""Test that replication_enable fails on driver raising exception."""
self.driver.do_setup(None)
replicated_volume = copy.deepcopy(REPLICATED_VOLUME)
# on purpose - set invalid value to replication_status
# expect an exception.
replicated_volume['replication_status'] = 'invalid_status_val'
self.assertRaises(
exception.CinderException,
self.driver.replication_enable,
CONTEXT,
replicated_volume
)
def test_list_replication_targets(self):
"""Return the list of replication targets for a volume."""
self.driver.do_setup(None)
expected_resp = {'targets': [{'target_device_id': 'fakedevice',
'type': 'managed'}],
'volume_id': VOLUME['id']}
targets = self.driver.list_replication_targets(CONTEXT, VOLUME)
self.assertEqual(expected_resp, targets)
def test_replication_failover(self):
"""Test that replication_failover returns successfully. """
self.driver.do_setup(None)
replicated_volume = copy.deepcopy(REPLICATED_VOLUME)
# assume the replication_status should be active
replicated_volume['replication_status'] = 'active' replicated_volume['replication_status'] = 'active'
expected_resp = {'host': FAKE_FAILOVER_HOST, expected_target_id = 'BLA'
'provider_location': FAKE_PROVIDER_LOCATION, expected_volume_update_list = [
'replication_driver_data': FAKE_DRIVER_DATA} {'volume_id': REPLICATED_VOLUME['id'],
'updates': {'replication_status': 'failed-over'}}]
model_update = self.driver.replication_failover( target_id, volume_update_list = self.driver.failover_host(
CONTEXT, CONTEXT,
replicated_volume, [replicated_volume],
SECONDARY SECONDARY
) )
self.assertEqual(expected_resp, model_update) self.assertEqual(expected_target_id, target_id)
self.assertEqual(expected_volume_update_list, volume_update_list)
def test_replication_failover_fail_on_cinder_exception(self): def test_failover_host_bad_state(self):
"""Test that replication_failover fails on CinderException. """ """Test that failover_host returns with error"""
self.driver.do_setup(None) self.driver.do_setup(None)
replicated_volume = copy.deepcopy(REPLICATED_VOLUME) replicated_volume = copy.deepcopy(REPLICATED_VOLUME)
# on purpose - set invalid value to replication_status # assume the replication_status is active
# expect an exception.
replicated_volume['replication_status'] = 'invalid_status_val' replicated_volume['replication_status'] = 'invalid_status_val'
self.assertRaises(
exception.CinderException, expected_target_id = 'BLA'
self.driver.replication_failover, expected_volume_update_list = [
{'volume_id': REPLICATED_VOLUME['id'],
'updates': {'replication_status': 'error'}}]
target_id, volume_update_list = self.driver.failover_host(
CONTEXT, CONTEXT,
replicated_volume, [replicated_volume],
SECONDARY SECONDARY
) )
self.assertEqual(expected_target_id, target_id)
self.assertEqual(expected_volume_update_list, volume_update_list)

View File

@ -63,7 +63,6 @@ class XIVDS8KDriver(san.SanDriver,
driver.ExtendVD, driver.ExtendVD,
driver.SnapshotVD, driver.SnapshotVD,
driver.MigrateVD, driver.MigrateVD,
driver.ReplicaVD,
driver.ConsistencyGroupVD, driver.ConsistencyGroupVD,
driver.CloneableImageVD, driver.CloneableImageVD,
driver.TransferVD): driver.TransferVD):
@ -78,6 +77,8 @@ class XIVDS8KDriver(san.SanDriver,
proxy = importutils.import_class(self.configuration.xiv_ds8k_proxy) proxy = importutils.import_class(self.configuration.xiv_ds8k_proxy)
active_backend_id = kwargs.get('active_backend_id', None)
# NOTE: All Array specific configurations are prefixed with: # NOTE: All Array specific configurations are prefixed with:
# "xiv_ds8k_array_" # "xiv_ds8k_array_"
# These additional flags should be specified in the cinder.conf # These additional flags should be specified in the cinder.conf
@ -96,7 +97,8 @@ class XIVDS8KDriver(san.SanDriver,
}, },
LOG, LOG,
exception, exception,
driver=self) driver=self,
active_backend_id=active_backend_id)
def do_setup(self, context): def do_setup(self, context):
"""Setup and verify IBM XIV and DS8K Storage connection.""" """Setup and verify IBM XIV and DS8K Storage connection."""
@ -217,26 +219,27 @@ class XIVDS8KDriver(san.SanDriver,
return self.xiv_ds8k_proxy.unmanage_volume(volume) return self.xiv_ds8k_proxy.unmanage_volume(volume)
def reenable_replication(self, context, volume): def freeze_backend(self, context):
"""Re-enable volume replication. """ """Notify the backend that it's frozen. """
return self.xiv_ds8k_proxy.reenable_replication(context, volume) return self.xiv_ds8k_proxy.freeze_backend(context)
def thaw_backend(self, context):
"""Notify the backend that it's unfrozen/thawed. """
return self.xiv_ds8k_proxy.thaw_backend(context)
def failover_host(self, context, volumes, secondary_id=None):
"""Failover a backend to a secondary replication target. """
return self.xiv_ds8k_proxy.failover_host(
context, volumes, secondary_id)
def get_replication_status(self, context, volume): def get_replication_status(self, context, volume):
"""Return replication status.""" """Return replication status."""
return self.xiv_ds8k_proxy.get_replication_status(context, volume) return self.xiv_ds8k_proxy.get_replication_status(context, volume)
def promote_replica(self, context, volume):
"""Promote the replica to be the primary volume."""
return self.xiv_ds8k_proxy.promote_replica(context, volume)
def create_replica_test_volume(self, volume, src_vref):
"""Creates a test replica clone of the specified replicated volume."""
return self.xiv_ds8k_proxy.create_replica_test_volume(volume, src_vref)
def retype(self, ctxt, volume, new_type, diff, host): def retype(self, ctxt, volume, new_type, diff, host):
"""Convert the volume to be of the new type.""" """Convert the volume to be of the new type."""
@ -280,24 +283,3 @@ class XIVDS8KDriver(san.SanDriver,
return self.xiv_ds8k_proxy.create_consistencygroup_from_src( return self.xiv_ds8k_proxy.create_consistencygroup_from_src(
context, group, volumes, cgsnapshot, snapshots, context, group, volumes, cgsnapshot, snapshots,
source_cg, source_vols) source_cg, source_vols)
def replication_disable(self, context, volume):
"""Disable replication on the specified volume."""
return self.xiv_ds8k_proxy.replication_disable(context, volume)
def replication_enable(self, context, volume):
"""Enable replication on a replication capable volume."""
return self.xiv_ds8k_proxy.replication_enable(context, volume)
def list_replication_targets(self, context, volume):
"""Provide a means to obtain replication targets for a volume."""
return self.xiv_ds8k_proxy.list_replication_targets(context, volume)
def replication_failover(self, context, volume, secondary):
"""Force failover to a secondary replication target. """
return self.xiv_ds8k_proxy.replication_failover(
context, volume, secondary)

View File

@ -0,0 +1,4 @@
---
features:
- Added replication v2.1 support to the IBM XIV/DS8K
driver.