This patch adds a new subgroup of commands called `util` with a single command ``remove_locks` that takes care of deleting locks that will no longer be used because the resources they are related to are no longer there. This works for volumes and snapshots and with the cinder services running or stopped. It gets the file locks filtering by name (must have the "cinder-" prefix and an UUID) and then if the service is online it will not delete file locks for resources that are still present in the database. Closes-Bug: #1432387 Change-Id: I2535017e112c8bcb9a2e516876f52a945e9c7da8
9.4 KiB
cinder-manage
Control and manage OpenStack block storage
- Author
- Copyright
-
OpenStack Foundation
- Manual section
-
1
- Manual group
-
cloud computing
SYNOPSIS
cinder-manage <category> <action> [<args>]
DESCRIPTION
cinder-manage
provides control of cinder database migration, and provides an interface
to get information about the current state of cinder. More information
about OpenStack Cinder is available at OpenStack
Cinder.
OPTIONS
The standard pattern for executing a cinder-manage command is:
cinder-manage <category> <command> [<args>]
For example, to obtain a list of the cinder services currently
running: cinder-manage service list
Run without arguments to see a list of available command categories:
cinder-manage
The categories are listed below, along with detailed descriptions.
You can also run with a category argument such as 'db' to see a list
of all commands in that category: cinder-manage db
These sections describe the available categories and arguments for cinder-manage.
Cinder Quota
Cinder quotas sometimes run out of sync, and while there are some mechanisms in place in Cinder that, with the proper configuration, try to do a resync of the quotas, they are not perfect and are susceptible to race conditions, so they may result in less than perfect accuracy in refreshed quotas.
The cinder-manage quota commands are meant to help manage these issues while allowing a finer control of when and what quotas are fixed.
Checking if quotas and reservations are correct.
cinder-manage quota check [-h] [--project-id PROJECT_ID] [--use-locks]
Accepted arguments are:
--project-id PROJECT_ID
The ID of the project where we want to sync the quotas
(defaults to all projects).
--use-locks For precise results tables in the DB need to be
locked.
This command checks quotas and reservations, for a specific project
(passing --project-id
) or for all projects, to see if they
are out of sync.
The check will also look for duplicated entries.
By default it runs in the least accurate mode (where races have a higher chance of happening) to minimize the impact on running cinder services. This means that false errors are more likely to be reported due to race conditions when Cinder services are running.
Accurate mode is also supported, but it will lock many tables (affecting all tenants) and is not recommended with services that are being used.
One way to use this action in combination with the sync action is to run the check for all projects, take note of those that are out of sync, and the sync them one by one at intervals to allow cinder to operate semi-normally.
Fixing quotas and reservations
cinder-manage quota sync [-h] [--project-id PROJECT_ID] [--no-locks]
Accepted arguments are:
--project-id PROJECT_ID
The ID of the project where we want to sync the quotas
(defaults to all projects).
--no-locks For less precise results, but also less intrusive.
This command refreshes existing quota usage and reservation count for a specific project or for all projects.
The refresh will also remove duplicated entries.
This operation is best executed when Cinder is not running, as it requires locking many tables (affecting all tenants) to make sure that then sync is accurate.
If accuracy is not our top priority, or we know that a specific project is not in use, we can disable the locking.
A different transaction is used for each project's quota sync, so an action failure will only rollback the current project's changes.
Cinder Db
cinder-manage db version
Print the current database version.
cinder-manage db sync [--bump-versions] [version]
Sync the database up to the most recent version. This is the standard way to create the db as well.
This command interprets the following options when it is invoked:
version Database version
- --bump-versions Update RPC and Objects versions when doing offline
-
upgrades, with this we no longer need to restart the services twice after the upgrade to prevent ServiceTooOld exceptions.
cinder-manage db purge [<number of days>]
Purge database entries that are marked as deleted, that are older than the number of days specified.
cinder-manage db online_data_migrations [--max_count <n>]
Perform online data migrations for database upgrade between releases in batches.
This command interprets the following options when it is invoked:
--max_count Maximum number of objects to migrate. If not specified, all
possible migrations will be completed, in batches of 50 at a
time.
Returns exit status 0 if no (further) updates are possible, 1 if the
--max_count
option was used and some updates were completed
successfully (even if others generated errors), 2 if some updates
generated errors and no other migrations were able to take effect in the
last batch attempted, or 127 if invalid input is provided (e.g.
non-numeric max-count).
This command should be run after upgrading the database schema. If it exits with partial updates (exit status 1) it should be called again, even if some updates initially generated errors, because some updates may depend on others having completed. If it exits with status 2, intervention is required to resolve the issue causing remaining updates to fail. It should be considered successfully completed only when the exit status is 0.
Cinder Logs
cinder-manage logs errors
Displays cinder errors from log files.
cinder-manage logs syslog [<number>]
Displays cinder the most recent entries from syslog. The optional number argument specifies the number of entries to display (default 10).
Cinder Volume
cinder-manage volume delete <volume_id>
Delete a volume without first checking that the volume is available.
cinder-manage volume update_host --currenthost <current host> --newhost <new host>
Updates the host name of all volumes currently associated with a specified host.
Cinder Host
cinder-manage host list [<zone>]
Displays a list of all physical hosts and their zone. The optional zone argument allows the list to be filtered on the requested zone.
Cinder Service
cinder-manage service list
Displays a list of all cinder services and their host, zone, status, state and when the information was last updated.
cinder-manage service remove <service> <host>
Removes a specified cinder service from a specified host.
Cinder Backup
cinder-manage backup list
Displays a list of all backups (including ones in progress) and the host on which the backup operation is running.
cinder-manage backup update_backup_host --currenthost <current host> --newhost <new host>
Updates the host name of all backups currently associated with a specified host.
Cinder Version
cinder-manage version list
Displays the codebase version cinder is running upon.
Cinder Config
cinder-manage config list [<param>]
Displays the current configuration parameters (options) for Cinder. The optional flag parameter may be used to display the configuration of one parameter.
Cinder Util
cinder-manage util clean_locks [-h] [--services-offline]
Clean file locks on the current host that were created and are used by drivers and cinder services for volumes, snapshots, and the backup service on the current host.
Should be run on any host where we are running a Cinder service (API, Scheduler, Volume, Backup) and can be run with the Cinder services running or stopped.
If the services are running it will check existing resources in the Cinder database in order to only remove resources that are no longer present (it's safe to delete the files).
For backups, the way to know if we can remove the startup lock is by checking if the PGRP in the file name is currently running cinder-backup.
Deleting locks while the services are offline is faster as there's no need to check the database or the running processes.
Default assumes that services are online, must pass
--services-offline
to specify that they are offline.
The common use case for running the command with
--services-offline
is to be called on startup as a service
unit before any cinder service is started. Command will be usually
called without the --services-offline
parameter manually or
from a cron job.
Warning
Passing --services-offline
when the Cinder services are
still running breaks the locking mechanism and can lead to undesired
behavior in ongoing Cinder operations.
Note
This command doesn't clean DLM locks (except when using file locks), as those don't leave lock leftovers.
FILES
The cinder.conf file contains configuration information in the form of python-gflags.
The cinder-manage.log file logs output from cinder-manage.
SEE ALSO
BUGS
- Cinder is hosted on Launchpad so you can view current bugs at Bugs : Cinder