Remove the Veritas HyperScale Driver
The Veritas HyperScale driver was marked unsupported in the Stein release. It hasn't reported on a patch in 327 days. This is outside Cinder's CI requirements and therefore the driver is being removed. Change-Id: I44a4dc52b2cced6dac98aa895aa3de1df0926fc5
This commit is contained in:
parent
0423642951
commit
9aca21f5ce
@ -1,522 +0,0 @@
|
||||
# Copyright (c) 2017 Veritas Technologies LLC. All rights reserved.
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
import mock
|
||||
|
||||
from cinder import context
|
||||
from cinder import exception
|
||||
from cinder import test
|
||||
from cinder.tests.unit import fake_snapshot
|
||||
from cinder.tests.unit import fake_volume
|
||||
from cinder.volume import configuration as conf
|
||||
from cinder.volume.drivers.veritas import exception as v_exception
|
||||
from cinder.volume.drivers.veritas import vrtshyperscale as vrts
|
||||
|
||||
|
||||
class FakeDb(object):
|
||||
def volume_metadata_get(self, *a, **kw):
|
||||
return {}
|
||||
|
||||
def volume_metadata_update(self, *a, **kw):
|
||||
return None
|
||||
|
||||
|
||||
def _stub_volume(*args, **kwargs):
|
||||
updates = {'provider_location': 'hyperscale-sv:/hyperscale'}
|
||||
return fake_volume.fake_db_volume(**updates)
|
||||
|
||||
|
||||
def _stub_snapshot(*args, **kwargs):
|
||||
updates = {'volume': _stub_volume(), 'name': 'vrts'}
|
||||
return fake_snapshot.fake_db_snapshot(**updates)
|
||||
|
||||
|
||||
def _stub_stats():
|
||||
data = {}
|
||||
data["volume_backend_name"] = 'Veritas_HyperScale'
|
||||
data["vendor_name"] = 'Veritas Technologies LLC'
|
||||
data["driver_version"] = '1.0'
|
||||
data["storage_protocol"] = 'nfs'
|
||||
data['total_capacity_gb'] = 0.0
|
||||
data['free_capacity_gb'] = 0.0
|
||||
data['reserved_percentage'] = 0
|
||||
data['QoS_support'] = False
|
||||
return data
|
||||
|
||||
|
||||
class VRTSHyperScaleDriverTestCase(test.TestCase):
|
||||
"""Test case for Veritas HyperScale VolumeDriver."""
|
||||
|
||||
driver_name = "cinder.volume.drivers.veritas.vrtshyperscale"
|
||||
|
||||
@staticmethod
|
||||
def gvmv_side_effect(arg1, arg2):
|
||||
"""Mock side effect for _get_volume_metadata_value."""
|
||||
# mock the return of get_volume_metadata_value
|
||||
# for different arguments
|
||||
if arg2 == 'Secondary_datanode_key':
|
||||
return '{9876}'
|
||||
elif arg2 == 'Secondary_datanode_ip':
|
||||
return '192.0.2.2'
|
||||
elif arg2 == 'current_dn_ip':
|
||||
return '192.0.2.1'
|
||||
elif arg2 == 'vsa_ip':
|
||||
return '192.0.2.1'
|
||||
|
||||
@mock.patch('cinder.volume.drivers.veritas.vrtshyperscale.HyperScaleDriver'
|
||||
'._fetch_config_for_compute')
|
||||
@mock.patch('cinder.volume.drivers.veritas.vrtshyperscale.HyperScaleDriver'
|
||||
'._fetch_config_for_datanode')
|
||||
@mock.patch('cinder.volume.drivers.veritas.vrtshyperscale.HyperScaleDriver'
|
||||
'._fetch_config_for_controller')
|
||||
def setUp(self, mock_fcfcntr, mock_fcfd, mock_fcfc):
|
||||
mock_fcfcntr.return_value = None
|
||||
mock_fcfd.return_value = None
|
||||
mock_fcfc.return_value = None
|
||||
|
||||
# Initialise a test seup
|
||||
super(VRTSHyperScaleDriverTestCase, self).setUp()
|
||||
|
||||
self.configuration = mock.Mock(conf.Configuration(None))
|
||||
self.configuration.reserved_percentage = 0
|
||||
self.context = context.get_admin_context()
|
||||
self.driver = vrts.HyperScaleDriver(
|
||||
db=FakeDb(), configuration=self.configuration)
|
||||
self.driver.dn_routing_key = '{1234}'
|
||||
self.driver.datanode_ip = '192.0.2.1'
|
||||
self.volume = _stub_volume()
|
||||
self.snapshot = _stub_snapshot()
|
||||
|
||||
@mock.patch('cinder.volume.drivers.veritas.vrtshyperscale.HyperScaleDriver'
|
||||
'._get_volume_metadata')
|
||||
@mock.patch('cinder.volume.drivers.veritas.vrtshyperscale.HyperScaleDriver'
|
||||
'._get_replicas')
|
||||
@mock.patch('cinder.volume.drivers.veritas.vrtshyperscale.HyperScaleDriver'
|
||||
'._get_volume_details_for_create_volume')
|
||||
@mock.patch('cinder.volume.drivers.veritas.utils'
|
||||
'.message_data_plane')
|
||||
@mock.patch('cinder.volume.api.API.update_volume_metadata')
|
||||
def test_create_volume_single_replicas(self, mock_uvm, mock_mdp,
|
||||
mock_gvdfcv, mock_get_replicas,
|
||||
mock_gvm):
|
||||
"""Test single volume replica. Happy path test case."""
|
||||
# Mock volume meatadata
|
||||
mock_gvm.return_value = _stub_volume()
|
||||
|
||||
# Mock number of replicas to 1
|
||||
mock_get_replicas.return_value = 1
|
||||
# assume volume details are populated correctly
|
||||
mock_gvdfcv.return_value = _stub_volume()
|
||||
|
||||
# assume volume message is sent to data node successfully
|
||||
mock_mdp.return_value = ("", None)
|
||||
# assume that the volume metadata gets updated correctly
|
||||
mock_uvm.return_value = {}
|
||||
|
||||
# declare the expected result
|
||||
expected_result = {
|
||||
'provider_location': 'hyperscale-sv:/hyperscale',
|
||||
'metadata': mock_gvm.return_value
|
||||
}
|
||||
|
||||
# call create volume and get the result
|
||||
actual_result = self.driver.create_volume(self.volume)
|
||||
|
||||
# Test if the return value matched the expected results
|
||||
self.assertDictEqual(actual_result, expected_result)
|
||||
|
||||
@mock.patch('cinder.volume.drivers.veritas.utils'
|
||||
'.get_hyperscale_version')
|
||||
def test_check_for_setup_error(self, mock_ghv):
|
||||
"""Test check for setup errors in Veritas HyperScale driver.
|
||||
|
||||
The test case checks happy path execution when driver version 1.0.0
|
||||
is installed.
|
||||
"""
|
||||
mock_ghv.return_value = "1.0.0"
|
||||
|
||||
# check the driver for setup errors
|
||||
self.driver.check_for_setup_error()
|
||||
|
||||
@mock.patch('cinder.volume.drivers.veritas.utils'
|
||||
'.get_hyperscale_version')
|
||||
def test_check_for_setup_error_unsupported_version(self, mock_ghv):
|
||||
"""Test check for setup errors in Veritas HyperScale driver.
|
||||
|
||||
The test case checks happy path execution when driver version 1.0.0
|
||||
is installed.
|
||||
"""
|
||||
mock_ghv.return_value = "1.0.0.1"
|
||||
|
||||
# check the driver for setup errors
|
||||
self.assertRaises(exception.VolumeBackendAPIException,
|
||||
self.driver.check_for_setup_error)
|
||||
|
||||
@mock.patch('cinder.volume.drivers.veritas.utils'
|
||||
'.get_hyperscale_version')
|
||||
def test_check_for_setup_error_exception(self, mock_ghv):
|
||||
"""Test check for setup errors in Veritas HyperScale driver.
|
||||
|
||||
The test case checks happy path execution when driver version 1.0.0
|
||||
is installed.
|
||||
"""
|
||||
mock_ghv.side_effect = v_exception.ErrorInHyperScaleVersion(
|
||||
cmd_error="mock error")
|
||||
|
||||
# check the driver for setup errors
|
||||
self.assertRaises(exception.VolumeBackendAPIException,
|
||||
self.driver.check_for_setup_error)
|
||||
|
||||
@mock.patch('cinder.volume.drivers.veritas.vrtshyperscale.HyperScaleDriver'
|
||||
'._get_volume_metadata_value')
|
||||
@mock.patch('cinder.volume.drivers.veritas.utils'
|
||||
'.message_data_plane')
|
||||
def test_delete_volume_no_replica(self, mock_mdp, mock_gvmv):
|
||||
"""Test happy path for delete_volume one data nodes."""
|
||||
mock_gvmv.return_value = None
|
||||
self.driver.delete_volume(self.volume)
|
||||
|
||||
message_body = {'display_name': self.volume['name']}
|
||||
|
||||
mock_mdp.assert_called_with(self.driver.dn_routing_key,
|
||||
'hyperscale.storage.dm.volume.delete',
|
||||
**message_body)
|
||||
|
||||
@mock.patch('cinder.volume.drivers.veritas.vrtshyperscale.HyperScaleDriver'
|
||||
'._get_volume_metadata_value')
|
||||
@mock.patch('cinder.volume.drivers.veritas.utils'
|
||||
'.message_data_plane')
|
||||
def test_delete_volume_more_than_one_replica(self, mock_mdp, mock_gvmv):
|
||||
"""Test happy path for delete_volume with more than one data nodes."""
|
||||
mock_gvmv.side_effect = VRTSHyperScaleDriverTestCase.gvmv_side_effect
|
||||
|
||||
message_body = {'display_name': self.volume['name']}
|
||||
|
||||
# make the delete call
|
||||
self.driver.delete_volume(self.volume)
|
||||
|
||||
# check if delete volume sent to reflection target on data node
|
||||
# check if mq message sent with 'Secondary_datanode_key'
|
||||
mock_mdp.assert_any_call(
|
||||
'{9876}', 'hyperscale.storage.dm.volume.delete', **message_body)
|
||||
|
||||
# check if the delete is sent to primary data node as well
|
||||
mock_mdp.assert_any_call(self.driver.dn_routing_key,
|
||||
'hyperscale.storage.dm.volume.delete',
|
||||
**message_body)
|
||||
|
||||
@mock.patch('cinder.volume.drivers.veritas.vrtshyperscale.HyperScaleDriver'
|
||||
'._get_volume_metadata_value')
|
||||
@mock.patch('cinder.volume.drivers.veritas.utils'
|
||||
'.message_data_plane')
|
||||
def test_delete_volume_no_replica_failure(self, mock_mdp, mock_gvmv):
|
||||
"""Failure case for delete_volume one node in data plane."""
|
||||
mock_gvmv.side_effect = None
|
||||
self.driver.delete_volume(self.volume)
|
||||
mock_mdp.side_effect = v_exception.UnableToProcessHyperScaleCmdOutput(
|
||||
cmd_out='mock error')
|
||||
self.assertRaises(exception.VolumeIsBusy, self.driver.delete_volume,
|
||||
self.volume)
|
||||
|
||||
@mock.patch('cinder.volume.drivers.veritas.vrtshyperscale.HyperScaleDriver'
|
||||
'._get_volume_metadata_value')
|
||||
@mock.patch('cinder.volume.drivers.veritas.utils'
|
||||
'.message_data_plane')
|
||||
def test_delete_volume_more_than_one_replica_failure(self, mock_mdp,
|
||||
mock_gvmv):
|
||||
"""failure case for delete_volume with more than one data nodes."""
|
||||
mock_gvmv.side_effect = VRTSHyperScaleDriverTestCase.gvmv_side_effect
|
||||
|
||||
mock_mdp.side_effect = v_exception.UnableToProcessHyperScaleCmdOutput(
|
||||
cmd_out='mock error')
|
||||
|
||||
self.assertRaises(exception.VolumeIsBusy, self.driver.delete_volume,
|
||||
self.volume)
|
||||
|
||||
@mock.patch('cinder.volume.drivers.veritas.utils'
|
||||
'.get_guid_with_curly_brackets')
|
||||
def test_delete_snapshot_force_flag(self, mock_ggwcb):
|
||||
"""Test snapshot deletion does not happen if force flag is set."""
|
||||
# get a mock snapshot object
|
||||
snapshot = fake_snapshot.fake_db_snapshot()
|
||||
# set the force in metadata of snapshot
|
||||
snapshot['metadata'] = {"force": "force"}
|
||||
|
||||
# call the delete volume
|
||||
self.driver.delete_snapshot(snapshot)
|
||||
|
||||
# if snapshot has force set in metadata then
|
||||
# get_guid_with_curly_brackets() will not be called because we
|
||||
# return as soon as we see force
|
||||
mock_ggwcb.assert_not_called()
|
||||
|
||||
def test_delete_snapshot_isbusy_flag(self):
|
||||
"""Test snapshot deletion throws exception if snapshot is busy."""
|
||||
# get a mock snapshot object
|
||||
snapshot = fake_snapshot.fake_db_snapshot()
|
||||
# set the force in metadata of snapshot
|
||||
snapshot['metadata'] = {"is_busy": "is_busy"}
|
||||
|
||||
# call the delete volume to check if it raises Busy Exception
|
||||
self.assertRaises(exception.SnapshotIsBusy,
|
||||
self.driver.delete_snapshot, snapshot)
|
||||
|
||||
@mock.patch('cinder.volume.drivers.veritas.vrtshyperscale.HyperScaleDriver'
|
||||
'._get_volume_metadata')
|
||||
@mock.patch('cinder.volume.api.API.get_volume')
|
||||
@mock.patch('cinder.volume.drivers.veritas.utils'
|
||||
'.message_data_plane')
|
||||
def test_delete_snapshot_from_primary_dn(self, mock_mdp, mock_gv,
|
||||
mock_gvm):
|
||||
"""Test snapshot deletion from primary DN."""
|
||||
# get mock volume
|
||||
mock_gv.return_value = None
|
||||
mock_gvm.return_value = {'current_dn_ip': self.driver.datanode_ip}
|
||||
|
||||
message_body = {}
|
||||
message_body['volume_guid'] = '{' + self.volume['id'] + '}'
|
||||
message_body['snapshot_id'] = '{' + self.snapshot['id'] + '}'
|
||||
|
||||
# call delete snapshot
|
||||
self.driver.delete_snapshot(self.snapshot)
|
||||
|
||||
# assert msg is sent over mq with primary DN routing key
|
||||
mock_mdp.assert_called_with(self.driver.dn_routing_key,
|
||||
'hyperscale.storage.dm.version.delete',
|
||||
**message_body)
|
||||
|
||||
@mock.patch('cinder.volume.drivers.veritas.vrtshyperscale.HyperScaleDriver'
|
||||
'._get_volume_metadata')
|
||||
@mock.patch('cinder.volume.api.API.get_volume')
|
||||
@mock.patch('cinder.volume.drivers.veritas.utils'
|
||||
'.message_data_plane')
|
||||
@mock.patch('cinder.volume.drivers.veritas.vrtshyperscale.HyperScaleDriver'
|
||||
'._get_volume_metadata_value')
|
||||
def test_delete_snapshot_from_current_dn(self, mock_gvmv, mock_mdp,
|
||||
mock_gv, mock_gvm):
|
||||
"""Test snapshot deletion DN value from volume."""
|
||||
# get a mock volume
|
||||
mock_gv.return_value = _stub_volume()
|
||||
|
||||
# get a mock value of DN from volume
|
||||
mock_gvmv.return_value = '{9876}'
|
||||
|
||||
message_body = {}
|
||||
message_body['volume_guid'] = '{' + self.volume['id'] + '}'
|
||||
message_body['snapshot_id'] = '{' + self.snapshot['id'] + '}'
|
||||
|
||||
# call delete snapshot
|
||||
self.driver.delete_snapshot(self.snapshot)
|
||||
|
||||
# assert msg is sent over mq with key from volume's current_dn_owner
|
||||
mock_mdp.assert_called_with(
|
||||
'{9876}', 'hyperscale.storage.dm.version.delete', **message_body)
|
||||
|
||||
@mock.patch('cinder.volume.drivers.veritas.utils'
|
||||
'.message_data_plane')
|
||||
def test_fetch_volume_stats_failure(self, mock_mdp):
|
||||
"""Test case checking failure of pool for fetching stats."""
|
||||
# since we have initialised the pool to None in setup()
|
||||
# the function will return only the stub without populating
|
||||
# any free and used stats
|
||||
mock_obj = {'payload': {}}
|
||||
|
||||
mock_mdp.return_value = (mock_obj, None)
|
||||
self.assertDictEqual(_stub_stats(), self.driver._fetch_volume_status())
|
||||
|
||||
@mock.patch('cinder.volume.drivers.veritas.utils'
|
||||
'.message_data_plane')
|
||||
def test_create_cloned_volume_with_exception(self, mock_mdp):
|
||||
"""Test case throws exception when command failed to execute."""
|
||||
vol_a = _stub_volume()
|
||||
vol_b = _stub_volume()
|
||||
mock_mdp.side_effect = v_exception.UnableToExecuteHyperScaleCmd(
|
||||
command='mock error')
|
||||
self.assertRaises(v_exception.UnableToExecuteHyperScaleCmd,
|
||||
self.driver.create_cloned_volume, vol_b, vol_a)
|
||||
|
||||
@mock.patch('cinder.volume.drivers.veritas.utils'
|
||||
'.message_data_plane')
|
||||
@mock.patch('cinder.volume.drivers.veritas.vrtshyperscale'
|
||||
'.HyperScaleDriver._select_rt')
|
||||
def test_create_cloned_volume_with_no_replica(self, mock_srt, mock_mdp):
|
||||
"""Test case clone volume when there is no replica."""
|
||||
mock_obj = {'payload': {}}
|
||||
mock_mdp.return_value = (mock_obj, None)
|
||||
mock_srt.return_value = (None, None)
|
||||
vol_a = _stub_volume()
|
||||
vol_b = _stub_volume()
|
||||
self.assertDictContainsSubset({
|
||||
'provider_location': 'hyperscale-sv:/hyperscale'
|
||||
}, self.driver.create_cloned_volume(vol_b, vol_a))
|
||||
|
||||
@mock.patch('cinder.volume.drivers.veritas.utils'
|
||||
'.message_data_plane')
|
||||
@mock.patch('cinder.volume.drivers.veritas.vrtshyperscale'
|
||||
'.HyperScaleDriver._select_rt')
|
||||
@mock.patch('cinder.volume.drivers.veritas.vrtshyperscale.HyperScaleDriver'
|
||||
'._get_volume_metadata_value')
|
||||
def test_create_cloned_volume_with_replica(self, mock_gvmv, mock_srt,
|
||||
mock_mdp):
|
||||
"""Test case clone volume when there is replica."""
|
||||
mock_gvmv.side_effect = VRTSHyperScaleDriverTestCase.gvmv_side_effect
|
||||
mock_obj = {'payload': {}}
|
||||
mock_mdp.return_value = (mock_obj, None)
|
||||
mock_srt.return_value = ('{1234}', '192.0.2.2')
|
||||
vol_a = _stub_volume()
|
||||
vol_b = _stub_volume()
|
||||
metadata = {
|
||||
'current_dn_owner': '{1234}',
|
||||
'Potential_secondary_key': '{1234}',
|
||||
'Primary_datanode_ip': '192.0.2.1',
|
||||
'Potential_secondary_ip': '192.0.2.2',
|
||||
'current_dn_ip': '192.0.2.1',
|
||||
'source_volid': vol_a['id'],
|
||||
'size': vol_a['size']
|
||||
}
|
||||
self.assertDictContainsSubset({
|
||||
'provider_location': 'hyperscale-sv:/hyperscale',
|
||||
'metadata': metadata
|
||||
}, self.driver.create_cloned_volume(vol_b, vol_a))
|
||||
|
||||
@mock.patch('cinder.volume.drivers.veritas.utils'
|
||||
'.message_data_plane')
|
||||
def test_extend_volume_with_exception(self, mock_mdp):
|
||||
"""Test case extend volume to the given size in GB."""
|
||||
mock_mdp.side_effect = v_exception.UnableToProcessHyperScaleCmdOutput(
|
||||
cmd_out='mock error')
|
||||
self.assertRaises(exception.VolumeDriverException,
|
||||
self.driver.extend_volume, _stub_volume(), 256)
|
||||
|
||||
@mock.patch('cinder.volume.drivers.veritas.utils'
|
||||
'.message_data_plane')
|
||||
def test_extend_volume_no_exception(self, mock_mdp):
|
||||
"""Test case extend volume thorws exception."""
|
||||
mock_mdp.return_value = (None, None)
|
||||
self.driver.extend_volume(_stub_volume(), 256)
|
||||
|
||||
@mock.patch('cinder.volume.drivers.veritas.utils'
|
||||
'.message_data_plane')
|
||||
def test_create_volume_from_snapshot_with_exception(self, mock_mdp):
|
||||
"""Test case create volume from snapshot thorws exception."""
|
||||
fake_volume, fake_snapshot = _stub_volume(), _stub_snapshot()
|
||||
mock_mdp.side_effect = v_exception.UnableToExecuteHyperScaleCmd(
|
||||
command='mock error')
|
||||
self.assertRaises(v_exception.UnableToExecuteHyperScaleCmd,
|
||||
self.driver.create_volume_from_snapshot, fake_volume,
|
||||
fake_snapshot)
|
||||
|
||||
@mock.patch('cinder.volume.drivers.veritas.utils'
|
||||
'.message_data_plane')
|
||||
@mock.patch('cinder.volume.drivers.veritas.vrtshyperscale'
|
||||
'.HyperScaleDriver._select_rt')
|
||||
def test_create_volume_from_snapshot_with_no_replica(self, mock_srt,
|
||||
mock_mdp):
|
||||
"""Test case create volume from snapshot when there is no replica."""
|
||||
mock_obj = {'payload': {}}
|
||||
mock_mdp.return_value = (mock_obj, None)
|
||||
mock_srt.return_value = (None, None)
|
||||
fake_volume, fake_snapshot = _stub_volume(), _stub_snapshot()
|
||||
self.assertDictContainsSubset({
|
||||
'provider_location': 'hyperscale-sv:/hyperscale'
|
||||
}, self.driver.create_volume_from_snapshot(fake_volume, fake_snapshot))
|
||||
|
||||
@mock.patch('cinder.volume.drivers.veritas.utils'
|
||||
'.message_data_plane')
|
||||
@mock.patch('cinder.volume.drivers.veritas.vrtshyperscale'
|
||||
'.HyperScaleDriver._select_rt')
|
||||
@mock.patch('cinder.volume.drivers.veritas.vrtshyperscale.HyperScaleDriver'
|
||||
'._get_volume_metadata_value')
|
||||
def test_create_volume_from_snapshot_with_replica(self, mock_gvmv,
|
||||
mock_srt, mock_mdp):
|
||||
"""Test case create volume from snapshot when there is replica."""
|
||||
mock_gvmv.side_effect = VRTSHyperScaleDriverTestCase.gvmv_side_effect
|
||||
mock_obj = {'payload': {}}
|
||||
mock_mdp.return_value = (mock_obj, None)
|
||||
mock_srt.return_value = ('{1234}', '192.0.2.2')
|
||||
fake_volume, fake_snapshot = _stub_volume(), _stub_snapshot()
|
||||
metadata = {
|
||||
'current_dn_owner': '{1234}',
|
||||
'Potential_secondary_key': '{1234}',
|
||||
'Primary_datanode_ip': '192.0.2.1',
|
||||
'Potential_secondary_ip': '192.0.2.2',
|
||||
'current_dn_ip': '192.0.2.1',
|
||||
'snapshot_id': fake_snapshot['id'],
|
||||
'parent_volume_guid': '{' + fake_snapshot['volume']['id'] + '}'
|
||||
}
|
||||
self.assertDictContainsSubset({
|
||||
'provider_location': 'hyperscale-sv:/hyperscale',
|
||||
'metadata': metadata
|
||||
}, self.driver.create_volume_from_snapshot(fake_volume, fake_snapshot))
|
||||
|
||||
def test_initialize_connection(self):
|
||||
"""Test case intialize_connection."""
|
||||
fake_volume = _stub_volume()
|
||||
expected_data = {
|
||||
'driver_volume_type': 'veritas_hyperscale',
|
||||
'data': {
|
||||
'export': fake_volume['provider_location'],
|
||||
'name': fake_volume['name']
|
||||
}
|
||||
}
|
||||
self.assertEqual(expected_data,
|
||||
self.driver.initialize_connection(fake_volume, None))
|
||||
|
||||
@mock.patch('cinder.volume.drivers.veritas.utils'
|
||||
'.message_compute_plane')
|
||||
@mock.patch('cinder.volume.drivers.veritas.utils'
|
||||
'.episodic_snap')
|
||||
@mock.patch('cinder.volume.drivers.veritas.vrtshyperscale.HyperScaleDriver'
|
||||
'._get_volume_metadata_value')
|
||||
def test_create_snapshot_with_exception(
|
||||
self, mock_gvmv, mock_es, mock_mcp):
|
||||
"""Test case create snapshot throws exception."""
|
||||
mock_gvmv.side_effect = VRTSHyperScaleDriverTestCase.gvmv_side_effect
|
||||
mock_es_obj = {'payload': {'update': False}}
|
||||
mock_es.return_value = mock_es_obj
|
||||
mock_mcp.side_effect = v_exception.UnableToExecuteHyperScaleCmd(
|
||||
command='mock error')
|
||||
fake_snapshot = _stub_snapshot()
|
||||
self.assertRaises(v_exception.UnableToExecuteHyperScaleCmd,
|
||||
self.driver.create_snapshot, fake_snapshot)
|
||||
|
||||
@mock.patch('cinder.volume.drivers.veritas.utils'
|
||||
'.message_controller')
|
||||
@mock.patch('cinder.volume.drivers.veritas.utils'
|
||||
'.message_data_plane')
|
||||
@mock.patch('cinder.volume.drivers.veritas.utils'
|
||||
'.episodic_snap')
|
||||
@mock.patch('cinder.volume.drivers.veritas.vrtshyperscale.HyperScaleDriver'
|
||||
'._get_volume_metadata_value')
|
||||
@mock.patch('cinder.volume.drivers.veritas.utils'
|
||||
'.message_compute_plane')
|
||||
def test_create_snapshot_user(
|
||||
self, mock_cdp, mock_gvmv, mock_es, mock_mdp, mock_mc):
|
||||
"""Test case user snapshot."""
|
||||
mock_gvmv.side_effect = VRTSHyperScaleDriverTestCase.gvmv_side_effect
|
||||
mock_es_obj = {'payload': {'update': False}}
|
||||
mock_es.return_value = mock_es_obj
|
||||
mock_obj = {'payload': {}}
|
||||
mock_mdp.return_value = ("", None)
|
||||
mock_mc.return_value = ("", None)
|
||||
mock_cdp.return_value = (mock_obj, None)
|
||||
fake_snapshot = _stub_snapshot()
|
||||
expected = {
|
||||
'metadata': {
|
||||
'status': 'creating',
|
||||
'datanode_ip': '192.0.2.1',
|
||||
'TYPE': vrts.TYPE_USER_SNAP
|
||||
}
|
||||
}
|
||||
self.assertEqual(expected, self.driver.create_snapshot(fake_snapshot))
|
@ -1,36 +0,0 @@
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
from cinder import exception
|
||||
from cinder.i18n import _
|
||||
|
||||
|
||||
class UnableToExecuteHyperScaleCmd(exception.VolumeDriverException):
|
||||
message = _("Failed HyperScale command for '%(command)s'")
|
||||
|
||||
|
||||
class UnableToProcessHyperScaleCmdOutput(exception.VolumeDriverException):
|
||||
message = _("Failed processing command output '%(cmd_out)s'"
|
||||
" for HyperScale command")
|
||||
|
||||
|
||||
class ErrorInFetchingConfiguration(exception.VolumeDriverException):
|
||||
message = _("Error in fetching configuration for '%(persona)s'")
|
||||
|
||||
|
||||
class ErrorInSendingMsg(exception.VolumeDriverException):
|
||||
message = _("Error in sending message '%(cmd_error)s'")
|
||||
|
||||
|
||||
class ErrorInHyperScaleVersion(exception.VolumeDriverException):
|
||||
message = _("Error in getting HyperScale version '%(cmd_error)s'")
|
@ -1,56 +0,0 @@
|
||||
# Copyright (c) 2017 Veritas Technologies LLC. All rights reserved.
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
"""
|
||||
Error Codes
|
||||
"""
|
||||
|
||||
EX_BAD_PARAM = 10
|
||||
EX_BAD_MESSAGE = 106
|
||||
MSG_SUCCESS = 0
|
||||
MSG_ERROR = 1
|
||||
|
||||
"""
|
||||
Constants
|
||||
"""
|
||||
HS_VHOST = "/"
|
||||
ACK_YES = 1
|
||||
ACK_NO = 0
|
||||
BLK_YES = 1
|
||||
BLK_NO = 0
|
||||
EXCH_DIRECT = "direct"
|
||||
EXCH_FANOUT = "fanout"
|
||||
EXCH_TOPIC = "topic"
|
||||
|
||||
MSG_REQUEST = 1
|
||||
MSG_RESPONSE = 2
|
||||
MSG_TOKEN = "token"
|
||||
MSG_OWNER = "owner"
|
||||
MSG_TYPE = "type"
|
||||
MSG_ERROR = "err_code"
|
||||
MSG_ACK = "ack"
|
||||
MSG_BLK = "blocking"
|
||||
MSG_BLK_INFO = "blocking_info"
|
||||
MSG_BLK_NAME = "name"
|
||||
MSG_BLK_BINDKEY = "bindkey"
|
||||
MSG_BLK_TYPE = "type"
|
||||
MSG_PAYLOAD = "payload"
|
||||
|
||||
# HyperScale Controller Exchange
|
||||
HS_CONTROLLER_EXCH = 'hyperscale-controller'
|
||||
HS_RPC_EXCH = 'hyperscale-recv'
|
||||
HS_DATANODE_EXCH = 'hyperscale-datanode'
|
||||
HS_COMPUTE_EXCH = 'hyperscale-storage'
|
||||
|
||||
SNAP_RESTORE_RF = 3
|
@ -1,352 +0,0 @@
|
||||
# Copyright (c) 2017 Veritas Technologies LLC. All rights reserved.
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
import json
|
||||
import uuid
|
||||
|
||||
from oslo_log import log as logging
|
||||
from oslo_utils import excutils
|
||||
import six
|
||||
|
||||
from cinder.privsep import hscli
|
||||
from cinder.volume.drivers.veritas import exception
|
||||
from cinder.volume.drivers.veritas import hs_constants as constants
|
||||
|
||||
LOG = logging.getLogger(__name__)
|
||||
|
||||
|
||||
def _populate_message_body(kwargs):
|
||||
message_body = {}
|
||||
# Build message body from kwargs
|
||||
for key, value in kwargs.items():
|
||||
if value is not None:
|
||||
message_body[key] = value
|
||||
|
||||
return message_body
|
||||
|
||||
|
||||
def generate_routingkey():
|
||||
return six.text_type(uuid.uuid1())
|
||||
|
||||
|
||||
def get_guid_with_curly_brackets(guid):
|
||||
return "{%s}" % guid if guid else guid
|
||||
|
||||
|
||||
def get_hyperscale_image_id():
|
||||
return "{%s}" % uuid.uuid1()
|
||||
|
||||
|
||||
def get_hyperscale_version():
|
||||
|
||||
version = None
|
||||
cmd_err = None
|
||||
try:
|
||||
cmd_arg = {'operation': 'version'}
|
||||
# create a json for cmd argument
|
||||
cmdarg_json = json.dumps(cmd_arg)
|
||||
|
||||
# call hscli for version
|
||||
(cmd_out, cmd_err) = hscli.hsexecute(cmdarg_json)
|
||||
|
||||
# cmd_err should be None in case of successful execution of cmd
|
||||
if not cmd_err:
|
||||
processed_output = process_cmd_out(cmd_out)
|
||||
version = processed_output.get('payload')
|
||||
else:
|
||||
LOG.error("Error %s in getting hyperscale version",
|
||||
cmd_err)
|
||||
raise exception.ErrorInHyperScaleVersion(cmd_err=cmd_err)
|
||||
except (exception.UnableToExecuteHyperScaleCmd,
|
||||
exception.UnableToProcessHyperScaleCmdOutput):
|
||||
LOG.error("Exception in running the command for version",
|
||||
exc_info=True)
|
||||
raise exception.UnableToExecuteHyperScaleCmd(command="version")
|
||||
|
||||
return version
|
||||
|
||||
|
||||
def get_datanode_id():
|
||||
|
||||
dnid = None
|
||||
cmd_out = None
|
||||
cmd_err = None
|
||||
try:
|
||||
cmd_arg = {'operation': 'get_datanode_id'}
|
||||
# create a json for cmd argument
|
||||
cmdarg_json = json.dumps(cmd_arg)
|
||||
|
||||
# call hscli for get_datanode_id
|
||||
(cmd_out, cmd_err) = hscli.hsexecute(cmdarg_json)
|
||||
|
||||
# cmd_err should be None in case of successful execution of cmd
|
||||
if not cmd_err:
|
||||
processed_output = process_cmd_out(cmd_out)
|
||||
dnid = processed_output.get('payload')
|
||||
else:
|
||||
LOG.error("Error %s in getting datanode hypervisor id",
|
||||
cmd_err)
|
||||
raise exception.UnableToExecuteHyperScaleCmd(
|
||||
command=cmdarg_json)
|
||||
except exception.UnableToExecuteHyperScaleCmd:
|
||||
with excutils.save_and_reraise_exception():
|
||||
LOG.debug("Unable to execute get_datanode_id", exc_info=True)
|
||||
|
||||
except exception.UnableToProcessHyperScaleCmdOutput:
|
||||
with excutils.save_and_reraise_exception():
|
||||
LOG.debug("Unable to process get_datanode_id output",
|
||||
exc_info=True)
|
||||
return dnid
|
||||
|
||||
|
||||
def episodic_snap(meta):
|
||||
|
||||
cmd_out = None
|
||||
cmd_err = None
|
||||
out_meta = None
|
||||
try:
|
||||
cmd_arg = {}
|
||||
cmd_arg['operation'] = 'episodic_snap'
|
||||
cmd_arg['metadata'] = meta
|
||||
# create a json for cmd argument
|
||||
cmdarg_json = json.dumps(cmd_arg)
|
||||
|
||||
# call hscli for episodic_snap
|
||||
(cmd_out, cmd_err) = hscli.hsexecute(cmdarg_json)
|
||||
|
||||
# cmd_err should be None in case of successful execution of cmd
|
||||
if not cmd_err:
|
||||
processed_output = process_cmd_out(cmd_out)
|
||||
out_meta = processed_output.get('payload')
|
||||
else:
|
||||
LOG.error("Error %s in processing episodic_snap",
|
||||
cmd_err)
|
||||
raise exception.UnableToExecuteHyperScaleCmd(
|
||||
command=cmdarg_json)
|
||||
except exception.UnableToExecuteHyperScaleCmd:
|
||||
with excutils.save_and_reraise_exception():
|
||||
LOG.debug("Unable to execute episodic_snap", exc_info=True)
|
||||
|
||||
except exception.UnableToProcessHyperScaleCmdOutput:
|
||||
with excutils.save_and_reraise_exception():
|
||||
LOG.debug("Unable to process episodic_snap output",
|
||||
exc_info=True)
|
||||
return out_meta
|
||||
|
||||
|
||||
def get_image_path(image_id, op_type='image'):
|
||||
|
||||
cmd_out = None
|
||||
cmd_err = None
|
||||
image_path = None
|
||||
try:
|
||||
cmd_arg = {}
|
||||
if op_type == 'image':
|
||||
cmd_arg['operation'] = 'get_image_path'
|
||||
elif op_type == 'volume':
|
||||
cmd_arg['operation'] = 'get_volume_path'
|
||||
cmd_arg['image_id'] = image_id
|
||||
# create a json for cmd argument
|
||||
cmdarg_json = json.dumps(cmd_arg)
|
||||
|
||||
# call hscli for get_image_path
|
||||
(cmd_out, cmd_err) = hscli.hsexecute(cmdarg_json)
|
||||
|
||||
# cmd_err should be None in case of successful execution of cmd
|
||||
if not cmd_err:
|
||||
processed_output = process_cmd_out(cmd_out)
|
||||
image_path = processed_output.get('payload')
|
||||
else:
|
||||
LOG.error("Error %s in processing get_image_path",
|
||||
cmd_err)
|
||||
raise exception.UnableToExecuteHyperScaleCmd(
|
||||
command=cmdarg_json)
|
||||
except exception.UnableToExecuteHyperScaleCmd:
|
||||
with excutils.save_and_reraise_exception():
|
||||
LOG.debug("Unable to execute get_image_path", exc_info=True)
|
||||
|
||||
except exception.UnableToProcessHyperScaleCmdOutput:
|
||||
with excutils.save_and_reraise_exception():
|
||||
LOG.debug("Unable to process get_image_path output",
|
||||
exc_info=True)
|
||||
return image_path
|
||||
|
||||
|
||||
def update_image(image_path, volume_id, hs_img_id):
|
||||
cmd_out = None
|
||||
cmd_err = None
|
||||
output = None
|
||||
try:
|
||||
cmd_arg = {}
|
||||
cmd_arg['operation'] = 'update_image'
|
||||
cmd_arg['image_path'] = image_path
|
||||
cmd_arg['volume_id'] = volume_id
|
||||
cmd_arg['hs_image_id'] = hs_img_id
|
||||
# create a json for cmd argument
|
||||
cmdarg_json = json.dumps(cmd_arg)
|
||||
|
||||
(cmd_out, cmd_err) = hscli.hsexecute(cmdarg_json)
|
||||
|
||||
# cmd_err should be None in case of successful execution of cmd
|
||||
if not cmd_err:
|
||||
output = process_cmd_out(cmd_out)
|
||||
else:
|
||||
LOG.error("Error %s in execution of update_image",
|
||||
cmd_err)
|
||||
raise exception.UnableToExecuteHyperScaleCmd(
|
||||
command=cmdarg_json)
|
||||
except exception.UnableToExecuteHyperScaleCmd:
|
||||
with excutils.save_and_reraise_exception():
|
||||
LOG.debug("Unable to execute update_image", exc_info=True)
|
||||
|
||||
except exception.UnableToProcessHyperScaleCmdOutput:
|
||||
with excutils.save_and_reraise_exception():
|
||||
LOG.debug("Unable to process update_image output",
|
||||
exc_info=True)
|
||||
return output
|
||||
|
||||
|
||||
def process_cmd_out(cmd_out):
|
||||
"""Process the cmd output."""
|
||||
|
||||
output = None
|
||||
|
||||
try:
|
||||
# get the python object from the cmd_out
|
||||
output = json.loads(cmd_out)
|
||||
error_code = output.get('err_code')
|
||||
if error_code:
|
||||
error_message = output.get('err_msg')
|
||||
operation = output.get('token')
|
||||
LOG.error("Failed to perform %(operation)s with error code"
|
||||
" %(err_code)s, error message is %(err_msg)s",
|
||||
{"operation": operation,
|
||||
"err_code": error_code,
|
||||
"err_msg": error_message})
|
||||
except ValueError:
|
||||
raise exception.UnableToProcessHyperScaleCmdOutput(cmd_out=cmd_out)
|
||||
|
||||
return output
|
||||
|
||||
|
||||
def check_for_setup_error():
|
||||
return True
|
||||
|
||||
|
||||
def get_configuration(persona):
|
||||
"""Get required configuration from controller."""
|
||||
|
||||
msg_body = {'persona': persona}
|
||||
configuration = None
|
||||
try:
|
||||
cmd_out, cmd_error = message_controller(
|
||||
constants.HS_CONTROLLER_EXCH,
|
||||
'hyperscale.controller.get.configuration',
|
||||
**msg_body)
|
||||
LOG.debug("Response Message from Controller: %s", cmd_out)
|
||||
payload = cmd_out.get('payload')
|
||||
configuration = payload.get('config_data')
|
||||
|
||||
except (exception.ErrorInSendingMsg,
|
||||
exception.UnableToExecuteHyperScaleCmd,
|
||||
exception.UnableToProcessHyperScaleCmdOutput):
|
||||
LOG.exception("Failed to get configuration from controller")
|
||||
raise exception.ErrorInFetchingConfiguration(persona=persona)
|
||||
|
||||
return configuration
|
||||
|
||||
|
||||
def _send_message(exchange, routing_key, message_token, **kwargs):
|
||||
"""Send message to specified node."""
|
||||
|
||||
cmd_out = None
|
||||
cmd_err = None
|
||||
processed_output = None
|
||||
msg = None
|
||||
try:
|
||||
LOG.debug("Sending message: %s", message_token)
|
||||
|
||||
# Build message from kwargs
|
||||
message_body = _populate_message_body(kwargs)
|
||||
cmd_arg = {}
|
||||
cmd_arg["operation"] = "message"
|
||||
cmd_arg["msg_body"] = message_body
|
||||
cmd_arg["msg_token"] = message_token
|
||||
# exchange name
|
||||
cmd_arg["exchange_name"] = exchange
|
||||
# routing key
|
||||
cmd_arg["routing_key"] = routing_key
|
||||
# create a json for cmd argument
|
||||
cmdarg_json = json.dumps(cmd_arg)
|
||||
|
||||
(cmd_out, cmd_err) = hscli.hsexecute(cmdarg_json)
|
||||
|
||||
# cmd_err should be none in case of successful execution of cmd
|
||||
if cmd_err:
|
||||
LOG.debug("Sending message failed. Error %s", cmd_err)
|
||||
raise exception.ErrorInSendingMsg(cmd_err=cmd_err)
|
||||
else:
|
||||
processed_output = process_cmd_out(cmd_out)
|
||||
|
||||
except exception.UnableToExecuteHyperScaleCmd:
|
||||
with excutils.save_and_reraise_exception():
|
||||
msg = ("Unable to execute HyperScale command for %(cmd)s"
|
||||
" to exchange %(exch)s with key %(rt_key)s")
|
||||
LOG.debug(msg, {"cmd": message_token,
|
||||
"exch": exchange,
|
||||
"rt_key": routing_key},
|
||||
exc_info=True)
|
||||
|
||||
except exception.UnableToProcessHyperScaleCmdOutput:
|
||||
with excutils.save_and_reraise_exception():
|
||||
msg = ("Unable to process msg %(message)s"
|
||||
" to exchange %(exch)s with key %(rt_key)s")
|
||||
LOG.debug(msg, {"message": message_token,
|
||||
"exch": exchange,
|
||||
"rt_key": routing_key})
|
||||
|
||||
return (processed_output, cmd_err)
|
||||
|
||||
|
||||
def message_compute_plane(routing_key, message_token, **kwargs):
|
||||
"""Send message to compute plane."""
|
||||
|
||||
LOG.debug("Sending message to compute plane")
|
||||
|
||||
return _send_message(constants.HS_COMPUTE_EXCH,
|
||||
routing_key,
|
||||
message_token,
|
||||
**kwargs)
|
||||
|
||||
|
||||
def message_data_plane(routing_key, message_token, **kwargs):
|
||||
"""Send message to data node."""
|
||||
|
||||
LOG.debug("Sending message to data plane")
|
||||
|
||||
return _send_message(constants.HS_DATANODE_EXCH,
|
||||
routing_key,
|
||||
message_token,
|
||||
**kwargs)
|
||||
|
||||
|
||||
def message_controller(routing_key, message_token, **kwargs):
|
||||
"""Send message to controller."""
|
||||
|
||||
LOG.debug("Sending message to controller")
|
||||
|
||||
return _send_message(constants.HS_CONTROLLER_EXCH,
|
||||
routing_key,
|
||||
message_token,
|
||||
**kwargs)
|
File diff suppressed because it is too large
Load Diff
@ -165,9 +165,6 @@ title=StorPool Storage Driver (storpool)
|
||||
[driver.synology]
|
||||
title=Synology Storage Driver (iSCSI)
|
||||
|
||||
[driver.vrtshyperscale]
|
||||
title=Veritas HyperScale Driver (veritas)
|
||||
|
||||
[driver.vrtsaccess]
|
||||
title=Veritas Access iSCSI Driver (iSCSI)
|
||||
|
||||
@ -249,8 +246,7 @@ driver.rbd=complete
|
||||
driver.sheepdog=complete
|
||||
driver.storpool=complete
|
||||
driver.synology=complete
|
||||
driver.vrtshyperscale=complete
|
||||
driver.vrtsaccess=missing
|
||||
driver.vrtsaccess=complete
|
||||
driver.vrtscnfs=complete
|
||||
driver.vzstorage=complete
|
||||
driver.vmware=complete
|
||||
@ -315,7 +311,6 @@ driver.storpool=complete
|
||||
driver.synology=complete
|
||||
driver.vrtsaccess=complete
|
||||
driver.vrtscnfs=complete
|
||||
driver.vrtshyperscale=missing
|
||||
driver.vzstorage=complete
|
||||
driver.vmware=complete
|
||||
driver.win_iscsi=complete
|
||||
@ -377,7 +372,6 @@ driver.rbd=missing
|
||||
driver.sheepdog=missing
|
||||
driver.storpool=missing
|
||||
driver.synology=missing
|
||||
driver.vrtshyperscale=missing
|
||||
driver.vrtsaccess=missing
|
||||
driver.vrtscnfs=missing
|
||||
driver.vzstorage=missing
|
||||
@ -444,7 +438,6 @@ driver.rbd=missing
|
||||
driver.sheepdog=missing
|
||||
driver.storpool=missing
|
||||
driver.synology=missing
|
||||
driver.vrtshyperscale=missing
|
||||
driver.vrtsaccess=missing
|
||||
driver.vrtscnfs=missing
|
||||
driver.vzstorage=missing
|
||||
@ -510,7 +503,6 @@ driver.rbd=complete
|
||||
driver.sheepdog=missing
|
||||
driver.storpool=complete
|
||||
driver.synology=missing
|
||||
driver.vrtshyperscale=missing
|
||||
driver.vrtsaccess=missing
|
||||
driver.vrtscnfs=missing
|
||||
driver.vzstorage=missing
|
||||
@ -577,7 +569,6 @@ driver.rbd=missing
|
||||
driver.sheepdog=missing
|
||||
driver.storpool=missing
|
||||
driver.synology=missing
|
||||
driver.vrtshyperscale=missing
|
||||
driver.vrtsaccess=missing
|
||||
driver.vrtscnfs=missing
|
||||
driver.vzstorage=missing
|
||||
@ -643,7 +634,6 @@ driver.rbd=complete
|
||||
driver.sheepdog=missing
|
||||
driver.storpool=missing
|
||||
driver.synology=missing
|
||||
driver.vrtshyperscale=missing
|
||||
driver.vrtsaccess=missing
|
||||
driver.vrtscnfs=missing
|
||||
driver.vzstorage=missing
|
||||
@ -710,7 +700,6 @@ driver.rbd=missing
|
||||
driver.sheepdog=missing
|
||||
driver.storpool=complete
|
||||
driver.synology=missing
|
||||
driver.vrtshyperscale=missing
|
||||
driver.vrtsaccess=missing
|
||||
driver.vrtscnfs=missing
|
||||
driver.vzstorage=missing
|
||||
@ -777,7 +766,6 @@ driver.rbd=missing
|
||||
driver.sheepdog=missing
|
||||
driver.storpool=missing
|
||||
driver.synology=missing
|
||||
driver.vrtshyperscale=missing
|
||||
driver.vrtsaccess=missing
|
||||
driver.vrtscnfs=missing
|
||||
driver.vzstorage=missing
|
||||
@ -841,7 +829,6 @@ driver.rbd=missing
|
||||
driver.sheepdog=missing
|
||||
driver.storpool=missing
|
||||
driver.synology=missing
|
||||
driver.vrtshyperscale=missing
|
||||
driver.vrtsaccess=missing
|
||||
driver.vrtscnfs=missing
|
||||
driver.vzstorage=missing
|
||||
|
@ -81,3 +81,4 @@ release.
|
||||
|
||||
* Train
|
||||
* Tintri Storage Driver
|
||||
* Veritas HyperScale Storage Driver
|
||||
|
@ -0,0 +1,15 @@
|
||||
---
|
||||
upgrade:
|
||||
- |
|
||||
The Veritas HyperScale storage driver has been removed after completion of its
|
||||
deprecation period without a reliable 3rd Party CI system being
|
||||
supported. Customers using the Veritas HyperScale driver should not upgrade
|
||||
Cinder without first migrating all volumes from their Veritas backend
|
||||
to a supported storage backend. Failure to migrate volumes will
|
||||
result in no longer being able to access volumes backed by the Veritas
|
||||
HyperScale storage backend.
|
||||
other:
|
||||
- |
|
||||
The Veritas HyperScale storage driver was marked unsupported in Stein due to
|
||||
3rd Party CI not meeting Cinder's requirements. As a result the
|
||||
driver is removed starting from the Train release.
|
Loading…
Reference in New Issue
Block a user