e6d4fe5f0f
There are cases where the timing between the start up for cinder-volume and cinder-scheduler service can result in a race where the scheduler doesn't know about the volume-service until the next periodic update. This change does attempts to do an easy fix by swapping the start order of the cinder services to ensure that the scheduler will be able to receive the volume service capabilities update. Fixes bug: 1189595 Change-Id: I8f477ddc04c15c04493f7ce6863e08e1de8f0128
514 lines
20 KiB
Plaintext
514 lines
20 KiB
Plaintext
# lib/cinder
|
||
# Install and start **Cinder** volume service
|
||
|
||
# Dependencies:
|
||
# - functions
|
||
# - DEST, DATA_DIR, STACK_USER must be defined
|
||
# SERVICE_{TENANT_NAME|PASSWORD} must be defined
|
||
# ``KEYSTONE_TOKEN_FORMAT`` must be defined
|
||
|
||
# stack.sh
|
||
# ---------
|
||
# install_cinder
|
||
# configure_cinder
|
||
# init_cinder
|
||
# start_cinder
|
||
# stop_cinder
|
||
# cleanup_cinder
|
||
|
||
# Save trace setting
|
||
XTRACE=$(set +o | grep xtrace)
|
||
set +o xtrace
|
||
|
||
|
||
# Defaults
|
||
# --------
|
||
|
||
# set up default driver
|
||
CINDER_DRIVER=${CINDER_DRIVER:-default}
|
||
|
||
# set up default directories
|
||
CINDER_DIR=$DEST/cinder
|
||
CINDERCLIENT_DIR=$DEST/python-cinderclient
|
||
CINDER_STATE_PATH=${CINDER_STATE_PATH:=$DATA_DIR/cinder}
|
||
CINDER_AUTH_CACHE_DIR=${CINDER_AUTH_CACHE_DIR:-/var/cache/cinder}
|
||
|
||
CINDER_CONF_DIR=/etc/cinder
|
||
CINDER_CONF=$CINDER_CONF_DIR/cinder.conf
|
||
CINDER_API_PASTE_INI=$CINDER_CONF_DIR/api-paste.ini
|
||
|
||
# Public facing bits
|
||
CINDER_SERVICE_HOST=${CINDER_SERVICE_HOST:-$SERVICE_HOST}
|
||
CINDER_SERVICE_PORT=${CINDER_SERVICE_PORT:-8776}
|
||
CINDER_SERVICE_PORT_INT=${CINDER_SERVICE_PORT_INT:-18776}
|
||
CINDER_SERVICE_PROTOCOL=${CINDER_SERVICE_PROTOCOL:-$SERVICE_PROTOCOL}
|
||
|
||
# Support entry points installation of console scripts
|
||
if [[ -d $CINDER_DIR/bin ]]; then
|
||
CINDER_BIN_DIR=$CINDER_DIR/bin
|
||
else
|
||
CINDER_BIN_DIR=$(get_python_exec_prefix)
|
||
fi
|
||
|
||
# Support for multi lvm backend configuration (default is no support)
|
||
CINDER_MULTI_LVM_BACKEND=$(trueorfalse False $CINDER_MULTI_LVM_BACKEND)
|
||
|
||
# Should cinder perform secure deletion of volumes?
|
||
# Defaults to true, can be set to False to avoid this bug when testing:
|
||
# https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1023755
|
||
CINDER_SECURE_DELETE=`trueorfalse True $CINDER_SECURE_DELETE`
|
||
|
||
# Cinder reports allocations back to the scheduler on periodic intervals
|
||
# it turns out we can get an "out of space" issue when we run tests too
|
||
# quickly just because cinder didn't realize we'd freed up resources.
|
||
# Make this configurable so that devstack-gate/tempest can set it to
|
||
# less than the 60 second default
|
||
# https://bugs.launchpad.net/cinder/+bug/1180976
|
||
CINDER_PERIODIC_INTERVAL=${CINDER_PERIODIC_INTERVAL:-60}
|
||
|
||
# Name of the lvm volume groups to use/create for iscsi volumes
|
||
VOLUME_GROUP=${VOLUME_GROUP:-stack-volumes}
|
||
VOLUME_BACKING_FILE=${VOLUME_BACKING_FILE:-$DATA_DIR/${VOLUME_GROUP}-backing-file}
|
||
VOLUME_BACKING_DEVICE=${VOLUME_BACKING_DEVICE:-}
|
||
|
||
# VOLUME_GROUP2 is used only if CINDER_MULTI_LVM_BACKEND = True
|
||
VOLUME_GROUP2=${VOLUME_GROUP2:-stack-volumes2}
|
||
VOLUME_BACKING_FILE2=${VOLUME_BACKING_FILE2:-$DATA_DIR/${VOLUME_GROUP2}-backing-file}
|
||
VOLUME_BACKING_DEVICE2=${VOLUME_BACKING_DEVICE2:-}
|
||
|
||
VOLUME_NAME_PREFIX=${VOLUME_NAME_PREFIX:-volume-}
|
||
|
||
|
||
# Functions
|
||
# ---------
|
||
# _clean_lvm_lv removes all cinder LVM volumes
|
||
# _clean_lvm_lv $VOLUME_GROUP $VOLUME_NAME_PREFIX
|
||
function _clean_lvm_lv() {
|
||
local vg=$1
|
||
local lv_prefix=$2
|
||
|
||
# Clean out existing volumes
|
||
for lv in `sudo lvs --noheadings -o lv_name $vg`; do
|
||
# lv_prefix prefixes the LVs we want
|
||
if [[ "${lv#$lv_prefix}" != "$lv" ]]; then
|
||
sudo lvremove -f $vg/$lv
|
||
fi
|
||
done
|
||
}
|
||
|
||
# _clean_lvm_backing_file() removes the backing file of the
|
||
# volume group used by cinder
|
||
# _clean_lvm_backing_file() $VOLUME_GROUP
|
||
function _clean_lvm_backing_file() {
|
||
local vg=$1
|
||
|
||
# if there is no logical volume left, it's safe to attempt a cleanup
|
||
# of the backing file
|
||
if [ -z "`sudo lvs --noheadings -o lv_name $vg`" ]; then
|
||
# if the backing physical device is a loop device, it was probably setup by devstack
|
||
VG_DEV=$(sudo losetup -j $DATA_DIR/${vg}-backing-file | awk -F':' '/backing-file/ { print $1}')
|
||
if [[ -n "$VG_DEV" ]]; then
|
||
sudo losetup -d $VG_DEV
|
||
rm -f $DATA_DIR/${vg}-backing-file
|
||
fi
|
||
fi
|
||
}
|
||
|
||
# cleanup_cinder() - Remove residual data files, anything left over from previous
|
||
# runs that a clean run would need to clean up
|
||
function cleanup_cinder() {
|
||
# ensure the volume group is cleared up because fails might
|
||
# leave dead volumes in the group
|
||
TARGETS=$(sudo tgtadm --op show --mode target)
|
||
if [ $? -ne 0 ]; then
|
||
# If tgt driver isn't running this won't work obviously
|
||
# So check the response and restart if need be
|
||
echo "tgtd seems to be in a bad state, restarting..."
|
||
if is_ubuntu; then
|
||
restart_service tgt
|
||
else
|
||
restart_service tgtd
|
||
fi
|
||
TARGETS=$(sudo tgtadm --op show --mode target)
|
||
fi
|
||
|
||
if [[ -n "$TARGETS" ]]; then
|
||
iqn_list=( $(grep --no-filename -r iqn $SCSI_PERSIST_DIR | sed 's/<target //' | sed 's/>//') )
|
||
for i in "${iqn_list[@]}"; do
|
||
echo removing iSCSI target: $i
|
||
sudo tgt-admin --delete $i
|
||
done
|
||
fi
|
||
|
||
if is_service_enabled cinder; then
|
||
sudo rm -rf $CINDER_STATE_PATH/volumes/*
|
||
fi
|
||
|
||
if is_ubuntu; then
|
||
stop_service tgt
|
||
else
|
||
stop_service tgtd
|
||
fi
|
||
|
||
# Campsite rule: leave behind a volume group at least as clean as we found it
|
||
_clean_lvm_lv $VOLUME_GROUP $VOLUME_NAME_PREFIX
|
||
_clean_lvm_backing_file $VOLUME_GROUP
|
||
|
||
if [ "$CINDER_MULTI_LVM_BACKEND" = "True" ]; then
|
||
_clean_lvm_lv $VOLUME_GROUP2 $VOLUME_NAME_PREFIX
|
||
_clean_lvm_backing_file $VOLUME_GROUP2
|
||
fi
|
||
}
|
||
|
||
# configure_cinder() - Set config files, create data dirs, etc
|
||
function configure_cinder() {
|
||
if [[ ! -d $CINDER_CONF_DIR ]]; then
|
||
sudo mkdir -p $CINDER_CONF_DIR
|
||
fi
|
||
sudo chown $STACK_USER $CINDER_CONF_DIR
|
||
|
||
cp -p $CINDER_DIR/etc/cinder/policy.json $CINDER_CONF_DIR
|
||
|
||
# Set the paths of certain binaries
|
||
CINDER_ROOTWRAP=$(get_rootwrap_location cinder)
|
||
|
||
# If Cinder ships the new rootwrap filters files, deploy them
|
||
# (owned by root) and add a parameter to $CINDER_ROOTWRAP
|
||
ROOTWRAP_CINDER_SUDOER_CMD="$CINDER_ROOTWRAP"
|
||
if [[ -d $CINDER_DIR/etc/cinder/rootwrap.d ]]; then
|
||
# Wipe any existing rootwrap.d files first
|
||
if [[ -d $CINDER_CONF_DIR/rootwrap.d ]]; then
|
||
sudo rm -rf $CINDER_CONF_DIR/rootwrap.d
|
||
fi
|
||
# Deploy filters to /etc/cinder/rootwrap.d
|
||
sudo mkdir -m 755 $CINDER_CONF_DIR/rootwrap.d
|
||
sudo cp $CINDER_DIR/etc/cinder/rootwrap.d/*.filters $CINDER_CONF_DIR/rootwrap.d
|
||
sudo chown -R root:root $CINDER_CONF_DIR/rootwrap.d
|
||
sudo chmod 644 $CINDER_CONF_DIR/rootwrap.d/*
|
||
# Set up rootwrap.conf, pointing to /etc/cinder/rootwrap.d
|
||
sudo cp $CINDER_DIR/etc/cinder/rootwrap.conf $CINDER_CONF_DIR/
|
||
sudo sed -e "s:^filters_path=.*$:filters_path=$CINDER_CONF_DIR/rootwrap.d:" -i $CINDER_CONF_DIR/rootwrap.conf
|
||
sudo chown root:root $CINDER_CONF_DIR/rootwrap.conf
|
||
sudo chmod 0644 $CINDER_CONF_DIR/rootwrap.conf
|
||
# Specify rootwrap.conf as first parameter to cinder-rootwrap
|
||
CINDER_ROOTWRAP="$CINDER_ROOTWRAP $CINDER_CONF_DIR/rootwrap.conf"
|
||
ROOTWRAP_CINDER_SUDOER_CMD="$CINDER_ROOTWRAP *"
|
||
fi
|
||
|
||
TEMPFILE=`mktemp`
|
||
echo "$USER ALL=(root) NOPASSWD: $ROOTWRAP_CINDER_SUDOER_CMD" >$TEMPFILE
|
||
chmod 0440 $TEMPFILE
|
||
sudo chown root:root $TEMPFILE
|
||
sudo mv $TEMPFILE /etc/sudoers.d/cinder-rootwrap
|
||
|
||
cp $CINDER_DIR/etc/cinder/api-paste.ini $CINDER_API_PASTE_INI
|
||
iniset $CINDER_API_PASTE_INI filter:authtoken auth_host $KEYSTONE_AUTH_HOST
|
||
iniset $CINDER_API_PASTE_INI filter:authtoken auth_port $KEYSTONE_AUTH_PORT
|
||
iniset $CINDER_API_PASTE_INI filter:authtoken auth_protocol $KEYSTONE_AUTH_PROTOCOL
|
||
iniset $CINDER_API_PASTE_INI filter:authtoken admin_tenant_name $SERVICE_TENANT_NAME
|
||
iniset $CINDER_API_PASTE_INI filter:authtoken admin_user cinder
|
||
iniset $CINDER_API_PASTE_INI filter:authtoken admin_password $SERVICE_PASSWORD
|
||
iniset $CINDER_API_PASTE_INI filter:authtoken signing_dir $CINDER_AUTH_CACHE_DIR
|
||
|
||
cp $CINDER_DIR/etc/cinder/cinder.conf.sample $CINDER_CONF
|
||
iniset $CINDER_CONF DEFAULT auth_strategy keystone
|
||
iniset $CINDER_CONF DEFAULT debug True
|
||
iniset $CINDER_CONF DEFAULT verbose True
|
||
if [ "$CINDER_MULTI_LVM_BACKEND" = "True" ]; then
|
||
iniset $CINDER_CONF DEFAULT enabled_backends lvmdriver-1,lvmdriver-2
|
||
iniset $CINDER_CONF lvmdriver-1 volume_group $VOLUME_GROUP
|
||
iniset $CINDER_CONF lvmdriver-1 volume_driver cinder.volume.drivers.lvm.LVMISCSIDriver
|
||
iniset $CINDER_CONF lvmdriver-1 volume_backend_name LVM_iSCSI
|
||
iniset $CINDER_CONF lvmdriver-2 volume_group $VOLUME_GROUP2
|
||
iniset $CINDER_CONF lvmdriver-2 volume_driver cinder.volume.drivers.lvm.LVMISCSIDriver
|
||
iniset $CINDER_CONF lvmdriver-2 volume_backend_name LVM_iSCSI_2
|
||
else
|
||
iniset $CINDER_CONF DEFAULT volume_group $VOLUME_GROUP
|
||
iniset $CINDER_CONF DEFAULT volume_name_template ${VOLUME_NAME_PREFIX}%s
|
||
fi
|
||
iniset $CINDER_CONF DEFAULT iscsi_helper tgtadm
|
||
iniset $CINDER_CONF DEFAULT sql_connection `database_connection_url cinder`
|
||
iniset $CINDER_CONF DEFAULT api_paste_config $CINDER_API_PASTE_INI
|
||
iniset $CINDER_CONF DEFAULT rootwrap_config "$CINDER_CONF_DIR/rootwrap.conf"
|
||
iniset $CINDER_CONF DEFAULT osapi_volume_extension cinder.api.contrib.standard_extensions
|
||
iniset $CINDER_CONF DEFAULT state_path $CINDER_STATE_PATH
|
||
iniset $CINDER_CONF DEFAULT periodic_interval $CINDER_PERIODIC_INTERVAL
|
||
|
||
if is_service_enabled tls-proxy; then
|
||
# Set the service port for a proxy to take the original
|
||
iniset $CINDER_CONF DEFAULT osapi_volume_listen_port $CINDER_SERVICE_PORT_INT
|
||
fi
|
||
|
||
if [ "$SYSLOG" != "False" ]; then
|
||
iniset $CINDER_CONF DEFAULT use_syslog True
|
||
fi
|
||
|
||
iniset_rpc_backend cinder $CINDER_CONF DEFAULT
|
||
|
||
if [[ "$CINDER_SECURE_DELETE" == "False" ]]; then
|
||
iniset $CINDER_CONF DEFAULT secure_delete False
|
||
iniset $CINDER_CONF DEFAULT volume_clear none
|
||
fi
|
||
|
||
if [ "$LOG_COLOR" == "True" ] && [ "$SYSLOG" == "False" ]; then
|
||
# Add color to logging output
|
||
iniset $CINDER_CONF DEFAULT logging_context_format_string "%(asctime)s.%(msecs)03d %(color)s%(levelname)s %(name)s [[01;36m%(request_id)s [00;36m%(user_id)s %(project_id)s%(color)s] [01;35m%(instance)s%(color)s%(message)s[00m"
|
||
iniset $CINDER_CONF DEFAULT logging_default_format_string "%(asctime)s.%(msecs)03d %(color)s%(levelname)s %(name)s [[00;36m-%(color)s] [01;35m%(instance)s%(color)s%(message)s[00m"
|
||
iniset $CINDER_CONF DEFAULT logging_debug_format_suffix "[00;33mfrom (pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d[00m"
|
||
iniset $CINDER_CONF DEFAULT logging_exception_prefix "%(color)s%(asctime)s.%(msecs)03d TRACE %(name)s [01;35m%(instance)s[00m"
|
||
fi
|
||
|
||
if [ "$CINDER_DRIVER" == "XenAPINFS" ]; then
|
||
(
|
||
set -u
|
||
iniset $CINDER_CONF DEFAULT volume_driver "cinder.volume.drivers.xenapi.sm.XenAPINFSDriver"
|
||
iniset $CINDER_CONF DEFAULT xenapi_connection_url "$CINDER_XENAPI_CONNECTION_URL"
|
||
iniset $CINDER_CONF DEFAULT xenapi_connection_username "$CINDER_XENAPI_CONNECTION_USERNAME"
|
||
iniset $CINDER_CONF DEFAULT xenapi_connection_password "$CINDER_XENAPI_CONNECTION_PASSWORD"
|
||
iniset $CINDER_CONF DEFAULT xenapi_nfs_server "$CINDER_XENAPI_NFS_SERVER"
|
||
iniset $CINDER_CONF DEFAULT xenapi_nfs_serverpath "$CINDER_XENAPI_NFS_SERVERPATH"
|
||
)
|
||
elif [ "$CINDER_DRIVER" == "sheepdog" ]; then
|
||
iniset $CINDER_CONF DEFAULT volume_driver "cinder.volume.drivers.sheepdog.SheepdogDriver"
|
||
elif [ "$CINDER_DRIVER" == "glusterfs" ]; then
|
||
# To use glusterfs, set the following in localrc:
|
||
# CINDER_DRIVER=glusterfs
|
||
# CINDER_GLUSTERFS_SHARES="127.0.0.1:/vol1;127.0.0.1:/vol2"
|
||
# Shares are <host>:<volume> and separated by semicolons.
|
||
|
||
iniset $CINDER_CONF DEFAULT volume_driver "cinder.volume.drivers.glusterfs.GlusterfsDriver"
|
||
iniset $CINDER_CONF DEFAULT glusterfs_shares_config "$CINDER_CONF_DIR/glusterfs_shares"
|
||
touch $CINDER_CONF_DIR/glusterfs_shares
|
||
if [ ! -z "$CINDER_GLUSTERFS_SHARES" ]; then
|
||
CINDER_GLUSTERFS_SHARES=$(echo $CINDER_GLUSTERFS_SHARES | tr ";" "\n")
|
||
echo "$CINDER_GLUSTERFS_SHARES" > $CINDER_CONF_DIR/glusterfs_shares
|
||
fi
|
||
fi
|
||
|
||
if [[ is_fedora && $DISTRO =~ (rhel6) ]]; then
|
||
# Cinder clones are slightly larger due to some extra
|
||
# metadata. RHEL6 will not allow auto-extending of LV's
|
||
# without this, leading to clones giving hard-to-track disk
|
||
# I/O errors.
|
||
# see https://bugzilla.redhat.com/show_bug.cgi?id=975052
|
||
sudo sed -i~ \
|
||
-e 's/snapshot_autoextend_threshold =.*/snapshot_autoextend_threshold = 80/' \
|
||
-e 's/snapshot_autoextend_percent =.*/snapshot_autoextend_percent = 20/' \
|
||
/etc/lvm/lvm.conf
|
||
fi
|
||
|
||
}
|
||
|
||
# create_cinder_accounts() - Set up common required cinder accounts
|
||
|
||
# Tenant User Roles
|
||
# ------------------------------------------------------------------
|
||
# service cinder admin # if enabled
|
||
|
||
# Migrated from keystone_data.sh
|
||
create_cinder_accounts() {
|
||
|
||
SERVICE_TENANT=$(keystone tenant-list | awk "/ $SERVICE_TENANT_NAME / { print \$2 }")
|
||
ADMIN_ROLE=$(keystone role-list | awk "/ admin / { print \$2 }")
|
||
|
||
# Cinder
|
||
if [[ "$ENABLED_SERVICES" =~ "c-api" ]]; then
|
||
CINDER_USER=$(keystone user-create \
|
||
--name=cinder \
|
||
--pass="$SERVICE_PASSWORD" \
|
||
--tenant_id $SERVICE_TENANT \
|
||
--email=cinder@example.com \
|
||
| grep " id " | get_field 2)
|
||
keystone user-role-add \
|
||
--tenant_id $SERVICE_TENANT \
|
||
--user_id $CINDER_USER \
|
||
--role_id $ADMIN_ROLE
|
||
if [[ "$KEYSTONE_CATALOG_BACKEND" = 'sql' ]]; then
|
||
CINDER_SERVICE=$(keystone service-create \
|
||
--name=cinder \
|
||
--type=volume \
|
||
--description="Cinder Volume Service" \
|
||
| grep " id " | get_field 2)
|
||
keystone endpoint-create \
|
||
--region RegionOne \
|
||
--service_id $CINDER_SERVICE \
|
||
--publicurl "$CINDER_SERVICE_PROTOCOL://$CINDER_SERVICE_HOST:$CINDER_SERVICE_PORT/v1/\$(tenant_id)s" \
|
||
--adminurl "$CINDER_SERVICE_PROTOCOL://$CINDER_SERVICE_HOST:$CINDER_SERVICE_PORT/v1/\$(tenant_id)s" \
|
||
--internalurl "$CINDER_SERVICE_PROTOCOL://$CINDER_SERVICE_HOST:$CINDER_SERVICE_PORT/v1/\$(tenant_id)s"
|
||
fi
|
||
fi
|
||
}
|
||
|
||
# create_cinder_cache_dir() - Part of the init_cinder() process
|
||
function create_cinder_cache_dir() {
|
||
# Create cache dir
|
||
sudo mkdir -p $CINDER_AUTH_CACHE_DIR
|
||
sudo chown $STACK_USER $CINDER_AUTH_CACHE_DIR
|
||
rm -f $CINDER_AUTH_CACHE_DIR/*
|
||
}
|
||
|
||
create_cinder_volume_group() {
|
||
# According to the ``CINDER_MULTI_LVM_BACKEND`` value, configure one or two default volumes
|
||
# group called ``stack-volumes`` (and ``stack-volumes2``) for the volume
|
||
# service if it (they) does (do) not yet exist. If you don't wish to use a
|
||
# file backed volume group, create your own volume group called ``stack-volumes``
|
||
# and ``stack-volumes2`` before invoking ``stack.sh``.
|
||
#
|
||
# The two backing files are ``VOLUME_BACKING_FILE_SIZE`` in size, and they are stored in
|
||
# the ``DATA_DIR``.
|
||
|
||
if ! sudo vgs $VOLUME_GROUP; then
|
||
if [ -z "$VOLUME_BACKING_DEVICE" ]; then
|
||
# Only create if the file doesn't already exists
|
||
[[ -f $VOLUME_BACKING_FILE ]] || truncate -s $VOLUME_BACKING_FILE_SIZE $VOLUME_BACKING_FILE
|
||
DEV=`sudo losetup -f --show $VOLUME_BACKING_FILE`
|
||
|
||
# Only create if the loopback device doesn't contain $VOLUME_GROUP
|
||
if ! sudo vgs $VOLUME_GROUP; then
|
||
sudo vgcreate $VOLUME_GROUP $DEV
|
||
fi
|
||
else
|
||
sudo vgcreate $VOLUME_GROUP $VOLUME_BACKING_DEVICE
|
||
fi
|
||
fi
|
||
if [ "$CINDER_MULTI_LVM_BACKEND" = "True" ]; then
|
||
#set up the second volume if CINDER_MULTI_LVM_BACKEND is enabled
|
||
|
||
if ! sudo vgs $VOLUME_GROUP2; then
|
||
if [ -z "$VOLUME_BACKING_DEVICE2" ]; then
|
||
# Only create if the file doesn't already exists
|
||
[[ -f $VOLUME_BACKING_FILE2 ]] || truncate -s $VOLUME_BACKING_FILE_SIZE $VOLUME_BACKING_FILE2
|
||
|
||
DEV=`sudo losetup -f --show $VOLUME_BACKING_FILE2`
|
||
|
||
# Only create if the loopback device doesn't contain $VOLUME_GROUP
|
||
if ! sudo vgs $VOLUME_GROUP2; then
|
||
sudo vgcreate $VOLUME_GROUP2 $DEV
|
||
fi
|
||
else
|
||
sudo vgcreate $VOLUME_GROUP2 $VOLUME_BACKING_DEVICE2
|
||
fi
|
||
fi
|
||
fi
|
||
|
||
mkdir -p $CINDER_STATE_PATH/volumes
|
||
}
|
||
|
||
# init_cinder() - Initialize database and volume group
|
||
function init_cinder() {
|
||
# Force nova volumes off
|
||
NOVA_ENABLED_APIS=$(echo $NOVA_ENABLED_APIS | sed "s/osapi_volume,//")
|
||
|
||
if is_service_enabled $DATABASE_BACKENDS; then
|
||
# (Re)create cinder database
|
||
recreate_database cinder utf8
|
||
|
||
# Migrate cinder database
|
||
$CINDER_BIN_DIR/cinder-manage db sync
|
||
fi
|
||
|
||
if is_service_enabled c-vol; then
|
||
|
||
create_cinder_volume_group
|
||
|
||
if sudo vgs $VOLUME_GROUP; then
|
||
if is_fedora || is_suse; then
|
||
# service is not started by default
|
||
start_service tgtd
|
||
fi
|
||
|
||
# Remove iscsi targets
|
||
sudo tgtadm --op show --mode target | grep $VOLUME_NAME_PREFIX | grep Target | cut -f3 -d ' ' | sudo xargs -n1 tgt-admin --delete || true
|
||
# Start with a clean volume group
|
||
_clean_lvm_lv $VOLUME_GROUP $VOLUME_NAME_PREFIX
|
||
if [ "$CINDER_MULTI_LVM_BACKEND" = "True" ]; then
|
||
_clean_lvm_lv $VOLUME_GROUP2 $VOLUME_NAME_PREFIX
|
||
fi
|
||
fi
|
||
fi
|
||
|
||
create_cinder_cache_dir
|
||
}
|
||
|
||
# install_cinder() - Collect source and prepare
|
||
function install_cinder() {
|
||
git_clone $CINDER_REPO $CINDER_DIR $CINDER_BRANCH
|
||
setup_develop $CINDER_DIR
|
||
}
|
||
|
||
# install_cinderclient() - Collect source and prepare
|
||
function install_cinderclient() {
|
||
git_clone $CINDERCLIENT_REPO $CINDERCLIENT_DIR $CINDERCLIENT_BRANCH
|
||
setup_develop $CINDERCLIENT_DIR
|
||
}
|
||
|
||
# apply config.d approach for cinder volumes directory
|
||
function _configure_tgt_for_config_d() {
|
||
if [[ ! -d /etc/tgt/stack.d/ ]]; then
|
||
sudo ln -sf $CINDER_STATE_PATH/volumes /etc/tgt/stack.d
|
||
echo "include /etc/tgt/stack.d/*" | sudo tee -a /etc/tgt/targets.conf
|
||
fi
|
||
}
|
||
|
||
# start_cinder() - Start running processes, including screen
|
||
function start_cinder() {
|
||
if is_service_enabled c-vol; then
|
||
# Delete any old stack.conf
|
||
sudo rm -f /etc/tgt/conf.d/stack.conf
|
||
_configure_tgt_for_config_d
|
||
if is_ubuntu; then
|
||
# tgt in oneiric doesn't restart properly if tgtd isn't running
|
||
# do it in two steps
|
||
sudo stop tgt || true
|
||
sudo start tgt
|
||
elif is_fedora; then
|
||
# bypass redirection to systemctl during restart
|
||
sudo /sbin/service --skip-redirect tgtd restart
|
||
elif is_suse; then
|
||
restart_service tgtd
|
||
else
|
||
# note for other distros: unstack.sh also uses the tgt/tgtd service
|
||
# name, and would need to be adjusted too
|
||
exit_distro_not_supported "restarting tgt"
|
||
fi
|
||
fi
|
||
|
||
screen_it c-api "cd $CINDER_DIR && $CINDER_BIN_DIR/cinder-api --config-file $CINDER_CONF"
|
||
screen_it c-sch "cd $CINDER_DIR && $CINDER_BIN_DIR/cinder-scheduler --config-file $CINDER_CONF"
|
||
screen_it c-bak "cd $CINDER_DIR && $CINDER_BIN_DIR/cinder-backup --config-file $CINDER_CONF"
|
||
screen_it c-vol "cd $CINDER_DIR && $CINDER_BIN_DIR/cinder-volume --config-file $CINDER_CONF"
|
||
|
||
# NOTE(jdg): For cinder, startup order matters. To ensure that repor_capabilities is received
|
||
# by the scheduler start the cinder-volume service last (or restart it) after the scheduler
|
||
# has started. This is a quick fix for lp bug/1189595
|
||
|
||
# Start proxies if enabled
|
||
if is_service_enabled c-api && is_service_enabled tls-proxy; then
|
||
start_tls_proxy '*' $CINDER_SERVICE_PORT $CINDER_SERVICE_HOST $CINDER_SERVICE_PORT_INT &
|
||
fi
|
||
}
|
||
|
||
# stop_cinder() - Stop running processes
|
||
function stop_cinder() {
|
||
# Kill the cinder screen windows
|
||
for serv in c-api c-bak c-sch c-vol; do
|
||
screen -S $SCREEN_NAME -p $serv -X kill
|
||
done
|
||
|
||
if is_service_enabled c-vol; then
|
||
if is_ubuntu; then
|
||
stop_service tgt
|
||
else
|
||
stop_service tgtd
|
||
fi
|
||
fi
|
||
}
|
||
|
||
|
||
# Restore xtrace
|
||
$XTRACE
|
||
|
||
# Local variables:
|
||
# mode: shell-script
|
||
# End:
|