# # Translators: msgid "" msgstr "" "Project-Id-Version: OpenStack Manuals\n" "POT-Creation-Date: 2014-01-17 07:14+0000\n" "PO-Revision-Date: 2014-01-16 20:25+0000\n" "Last-Translator: openstackjenkins \n" "Language-Team: Urdu (http://www.transifex.com/projects/p/openstack/language/ur/)\n" "MIME-Version: 1.0\n" "Content-Type: text/plain; charset=UTF-8\n" "Content-Transfer-Encoding: 8bit\n" "Language: ur\n" "Plural-Forms: nplurals=2; plural=(n != 1);\n" #: ./doc/config-reference/ch_imageservice.xml10(title) msgid "Image Service" msgstr "" #: ./doc/config-reference/ch_imageservice.xml11(para) msgid "" "Compute relies on an external image service to store virtual machine images " "and maintain a catalog of available images. By default, Compute is " "configured to use the OpenStack Image Service (Glance), which is currently " "the only supported image service." msgstr "" #: ./doc/config-reference/ch_imageservice.xml19(para) msgid "" "If your installation requires euca2ools to register new images, you must run" " the nova-objectstore service. " "This service provides an Amazon S3 front-end for Glance, which is required " "by euca2ools." msgstr "" #: ./doc/config-reference/ch_imageservice.xml26(para) msgid "" "You can modify many of the OpenStack Image Catalogue and Delivery Service. " "The following tables provide a comprehensive list." msgstr "" #: ./doc/config-reference/ch_config-overview.xml6(title) msgid "OpenStack configuration overview" msgstr "" #: ./doc/config-reference/ch_config-overview.xml7(para) msgid "" "OpenStack is a collection of open source project components that enable " "setting up cloud services. Each component uses similar configuration " "techniques and a common framework for INI file options." msgstr "" #: ./doc/config-reference/ch_config-overview.xml12(para) msgid "" "This guide pulls together multiple references and configuration options for " "the following OpenStack components:" msgstr "" #: ./doc/config-reference/ch_config-overview.xml16(para) msgid "OpenStack Identity" msgstr "" #: ./doc/config-reference/ch_config-overview.xml17(para) msgid "OpenStack Compute" msgstr "" #: ./doc/config-reference/ch_config-overview.xml18(para) msgid "OpenStack Image Service" msgstr "" #: ./doc/config-reference/ch_config-overview.xml19(para) msgid "OpenStack Networking" msgstr "" #: ./doc/config-reference/ch_config-overview.xml20(para) msgid "OpenStack Dashboard" msgstr "" #: ./doc/config-reference/ch_config-overview.xml21(para) msgid "OpenStack Object Storage" msgstr "" #: ./doc/config-reference/ch_config-overview.xml22(para) msgid "OpenStack Block Storage" msgstr "" #: ./doc/config-reference/ch_computeconfigure.xml6(title) msgid "Compute" msgstr "" #: ./doc/config-reference/ch_computeconfigure.xml7(para) msgid "" "The OpenStack Compute service is a cloud computing fabric controller, which " "is the main part of an IaaS system. You can use OpenStack Compute to host " "and manage cloud computing systems. This section describes the OpenStack " "Compute configuration options." msgstr "" #: ./doc/config-reference/ch_computeconfigure.xml12(para) msgid "" "To configure your Compute installation, you must define configuration " "options in these files:" msgstr "" #: ./doc/config-reference/ch_computeconfigure.xml16(para) msgid "" "nova.conf. Contains most of the Compute configuration " "options. Resides in the /etc/nova directory." msgstr "" #: ./doc/config-reference/ch_computeconfigure.xml21(para) msgid "" "api-paste.ini. Defines Compute limits. Resides in the " "/etc/nova directory." msgstr "" #: ./doc/config-reference/ch_computeconfigure.xml26(para) msgid "" "Related Image Service and Identity Service management configuration files." msgstr "" #: ./doc/config-reference/ch_computeconfigure.xml32(title) msgid "Configure logging" msgstr "" #: ./doc/config-reference/ch_computeconfigure.xml33(para) msgid "" "You can use nova.conf file to configure where Compute " "logs events, the level of logging, and log formats." msgstr "" #: ./doc/config-reference/ch_computeconfigure.xml36(para) msgid "" "To customize log formats for OpenStack Compute, use these configuration " "option settings." msgstr "" #: ./doc/config-reference/ch_computeconfigure.xml41(title) msgid "Configure authentication and authorization" msgstr "" #: ./doc/config-reference/ch_computeconfigure.xml42(para) msgid "" "There are different methods of authentication for the OpenStack Compute " "project, including no authentication. The preferred system is the OpenStack " "Identity Service, code-named Keystone." msgstr "" #: ./doc/config-reference/ch_computeconfigure.xml46(para) msgid "" "To customize authorization settings for Compute, see these configuration " "settings in nova.conf." msgstr "" #: ./doc/config-reference/ch_computeconfigure.xml49(para) msgid "" "To customize certificate authority settings for Compute, see these " "configuration settings in nova.conf." msgstr "" #: ./doc/config-reference/ch_computeconfigure.xml53(para) msgid "" "To customize Compute and the Identity service to use LDAP as a backend, " "refer to these configuration settings in nova.conf." msgstr "" #: ./doc/config-reference/ch_computeconfigure.xml62(title) msgid "Configure resize" msgstr "" #: ./doc/config-reference/ch_computeconfigure.xml63(para) msgid "" "Resize (or Server resize) is the ability to change the flavor of a server, " "thus allowing it to upscale or downscale according to user needs. For this " "feature to work properly, you might need to configure some underlying virt " "layers." msgstr "" #: ./doc/config-reference/ch_computeconfigure.xml68(title) #: ./doc/config-reference/compute/section_hypervisor_kvm.xml5(title) msgid "KVM" msgstr "" #: ./doc/config-reference/ch_computeconfigure.xml69(para) msgid "" "Resize on KVM is implemented currently by transferring the images between " "compute nodes over ssh. For KVM you need hostnames to resolve properly and " "passwordless ssh access between your compute hosts. Direct access from one " "compute host to another is needed to copy the VM file across." msgstr "" #: ./doc/config-reference/ch_computeconfigure.xml74(para) msgid "" "Cloud end users can find out how to resize a server by reading the OpenStack End User Guide." msgstr "" #: ./doc/config-reference/ch_computeconfigure.xml80(title) #: ./doc/config-reference/compute/section_compute-configure-migrations.xml289(title) msgid "XenServer" msgstr "" #: ./doc/config-reference/ch_computeconfigure.xml81(para) msgid "" "To get resize to work with XenServer (and XCP), you need to establish a root" " trust between all hypervisor nodes and provide an /image mount point to " "your hypervisors dom0." msgstr "" #: ./doc/config-reference/bk-config-ref.xml6(title) #: ./doc/config-reference/bk-config-ref.xml8(titleabbrev) msgid "OpenStack Configuration Reference" msgstr "" #: ./doc/config-reference/bk-config-ref.xml16(orgname) #: ./doc/config-reference/bk-config-ref.xml21(holder) msgid "OpenStack Foundation" msgstr "OpenStack فاؤنڈیشن , بنیاد" #: ./doc/config-reference/bk-config-ref.xml20(year) msgid "2013" msgstr "" #: ./doc/config-reference/bk-config-ref.xml23(productname) msgid "OpenStack" msgstr "OpenStack" #: ./doc/config-reference/bk-config-ref.xml24(releaseinfo) msgid "havana" msgstr "" #: ./doc/config-reference/bk-config-ref.xml28(remark) msgid "Copyright details are filled in by the template." msgstr "کاپی رائٹ کی تفصیلات سانچے کی طرف سے بھرے پڑے ہیں." #: ./doc/config-reference/bk-config-ref.xml33(para) msgid "" "This document is for system administrators who want to look up configuration" " options. It contains lists of configuration options available with " "OpenStack and uses auto-generation to generate options and the descriptions " "from the code for each project. It includes sample configuration files." msgstr "" #: ./doc/config-reference/bk-config-ref.xml43(date) msgid "2014-01-09" msgstr "" #: ./doc/config-reference/bk-config-ref.xml47(para) msgid "" "Removes content addressed in installation, merges duplicated content, and " "revises legacy references." msgstr "" #: ./doc/config-reference/bk-config-ref.xml55(date) msgid "2013-10-17" msgstr "" #: ./doc/config-reference/bk-config-ref.xml59(para) msgid "Havana release." msgstr "" #: ./doc/config-reference/bk-config-ref.xml65(date) msgid "2013-08-16" msgstr "" #: ./doc/config-reference/bk-config-ref.xml69(para) msgid "" "Moves Block Storage driver configuration information from the " "Block Storage Administration Guide to this reference." msgstr "" #: ./doc/config-reference/bk-config-ref.xml77(date) msgid "2013-06-10" msgstr "" #: ./doc/config-reference/bk-config-ref.xml81(para) msgid "Initial creation of Configuration Reference." msgstr "" #: ./doc/config-reference/ch_blockstorageconfigure.xml10(title) msgid "Block Storage" msgstr "" #: ./doc/config-reference/ch_blockstorageconfigure.xml11(para) msgid "" "The OpenStack Block Storage Service works with many different storage " "drivers that you can configure by using these instructions." msgstr "" #: ./doc/config-reference/ch_blockstorageconfigure.xml15(title) #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml192(title) msgid "cinder.conf configuration file" msgstr "" #: ./doc/config-reference/ch_blockstorageconfigure.xml16(para) msgid "" "The cinder.conf file is installed in " "/etc/cinder by default. When you manually install the " "Block Storage Service, the options in the cinder.conf " "file are set to default values." msgstr "" #: ./doc/config-reference/ch_blockstorageconfigure.xml20(para) msgid "This example shows a typical cinder.conf file:" msgstr "" #: ./doc/config-reference/ch_identityconfigure.xml10(title) msgid "Identity Service" msgstr "" #: ./doc/config-reference/ch_identityconfigure.xml11(para) msgid "" "This chapter details the OpenStack Identity Service configuration options. " "For installation prerequisites and step-by-step walkthroughs, see the " "OpenStack Installation Guide for your distribution " "(docs.openstack.org) and " "Cloud Administrator Guide." msgstr "" #: ./doc/config-reference/ch_objectstorageconfigure.xml6(title) msgid "Object Storage" msgstr "" #: ./doc/config-reference/ch_objectstorageconfigure.xml7(para) msgid "" "OpenStack Object Storage uses multiple configuration files for multiple " "services and background daemons, and to manage server " "configurations. Default configuration options appear in the " "[DEFAULT] section. You can override the default values by " "setting values in the other sections." msgstr "" #: ./doc/config-reference/ch_objectstorageconfigure.xml17(title) msgid "Object server configuration" msgstr "" #: ./doc/config-reference/ch_objectstorageconfigure.xml18(para) msgid "" "Find an example object server configuration at etc/object-server" ".conf-sample in the source code repository." msgstr "" #: ./doc/config-reference/ch_objectstorageconfigure.xml21(para) #: ./doc/config-reference/ch_objectstorageconfigure.xml48(para) #: ./doc/config-reference/ch_objectstorageconfigure.xml78(para) #: ./doc/config-reference/ch_objectstorageconfigure.xml105(para) msgid "The available configuration options are:" msgstr "" #: ./doc/config-reference/ch_objectstorageconfigure.xml39(title) msgid "Sample object server configuration file" msgstr "" #: ./doc/config-reference/ch_objectstorageconfigure.xml44(title) msgid "Container server configuration" msgstr "" #: ./doc/config-reference/ch_objectstorageconfigure.xml45(para) msgid "" "Find an example container server configuration at etc/container-" "server.conf-sample in the source code repository." msgstr "" #: ./doc/config-reference/ch_objectstorageconfigure.xml68(title) msgid "Sample container server configuration file" msgstr "" #: ./doc/config-reference/ch_objectstorageconfigure.xml74(title) msgid "Account server configuration" msgstr "" #: ./doc/config-reference/ch_objectstorageconfigure.xml75(para) msgid "" "Find an example account server configuration at etc/account-server" ".conf-sample in the source code repository." msgstr "" #: ./doc/config-reference/ch_objectstorageconfigure.xml96(title) msgid "Sample account server configuration file" msgstr "" #: ./doc/config-reference/ch_objectstorageconfigure.xml101(title) msgid "Proxy server configuration" msgstr "" #: ./doc/config-reference/ch_objectstorageconfigure.xml102(para) msgid "" "Find an example proxy server configuration at etc/proxy-server" ".conf-sample in the source code repository." msgstr "" #: ./doc/config-reference/ch_objectstorageconfigure.xml131(title) msgid "Sample proxy server configuration file" msgstr "" #: ./doc/config-reference/ch_dashboardconfigure.xml6(title) msgid "Dashboard" msgstr "" #: ./doc/config-reference/ch_dashboardconfigure.xml7(para) msgid "" "This chapter describes how to configure the OpenStack dashboard with Apache " "web server." msgstr "" #: ./doc/config-reference/ch_networkingconfigure.xml10(title) msgid "Networking" msgstr "" #: ./doc/config-reference/ch_networkingconfigure.xml11(para) msgid "" "This chapter explains the OpenStack Networking configuration options. For " "installation prerequisites, steps, and use cases, see the " "OpenStack Installation Guide for your distribution " "(docs.openstack.org) and " "Cloud Administrator Guide." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-xen.xml10(title) #: ./doc/config-reference/compute/section_compute-configure-xen.xml75(title) msgid "Xen configuration reference" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-xen.xml11(para) msgid "" "The following section discusses some commonly changed options in XenServer. " "The table below provides a complete reference of all configuration options " "available for configuring Xen with OpenStack." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-xen.xml15(para) msgid "" "The recommended way to use Xen with OpenStack is through the XenAPI driver. " "To enable the XenAPI driver, add the following configuration options " "/etc/nova/nova.conf and restart the nova-compute service:" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-xen.xml23(para) msgid "" "These connection details are used by the OpenStack Compute service to " "contact your hypervisor and are the same details you use to connect " "XenCenter, the XenServer management console, to your XenServer or XCP box." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-xen.xml27(para) msgid "" "The xenapi_connection_url is generally the management " "network IP address of the XenServer. Though it is possible to use the " "internal network IP Address (169.250.0.1) to contact XenAPI, this does not " "allow live migration between hosts, and other functionalities like host " "aggregates do not work." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-xen.xml34(para) msgid "" "It is possible to manage Xen using libvirt, though this is not well-tested " "or supported. To experiment using Xen through libvirt add the following " "configuration options /etc/nova/nova.conf: " "" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-xen.xml42(title) #: ./doc/config-reference/networking/section_networking-options-reference.xml19(title) msgid "Agent" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-xen.xml43(para) msgid "" "If you don't have the guest agent on your VMs, it takes a long time for nova" " to decide the VM has successfully started. Generally a large timeout is " "required for Windows instances, bug you may want to tweak " "agent_version_timeout" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-xen.xml48(title) msgid "Firewall" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-xen.xml49(para) msgid "" "If using nova-network, IPTables is supported: Alternately, " "doing the isolation in Dom0: " msgstr "" #: ./doc/config-reference/compute/section_compute-configure-xen.xml56(title) msgid "VNC proxy address" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-xen.xml57(para) msgid "" "Assuming you are talking to XenAPI through the host local management " "network, and XenServer is on the address: 169.254.0.1, you can use the " "following: vncserver_proxyclient_address=169.254.0.1" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-xen.xml63(title) msgid "Storage" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-xen.xml71(para) msgid "" "To use a XenServer pool, you must create the pool by using the Host " "Aggregates feature." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-xen.xml64(para) msgid "" "You can specify which Storage Repository to use with nova by looking at the " "following flag. The default is to use the local-storage setup by the default" " installer: Another good alternative is to use the " "\"default\" storage (for example if you have attached NFS or any other " "shared storage): " msgstr "" #. When image changes, this message will be marked fuzzy or untranslated for #. you. #. It doesn't matter what you translate it to: it's not used at all. #: ./doc/config-reference/compute/section_introduction-to-xen.xml127(None) msgid "" "@@image: '../../common/figures/xenserver_architecture.png'; " "md5=8eb25be1693aa7865967ac7b07d3e563" msgstr "" #: ./doc/config-reference/compute/section_introduction-to-xen.xml7(title) msgid "Xen, XenAPI, XenServer, and XCP" msgstr "" #: ./doc/config-reference/compute/section_introduction-to-xen.xml9(para) msgid "" "This section describes Xen, XenAPI, XenServer, and XCP, their differences, " "and how to use them with OpenStack. After you understand how the Xen and KVM" " architectures differ, you can determine when to use each architecture in " "your OpenStack cloud." msgstr "" #: ./doc/config-reference/compute/section_introduction-to-xen.xml15(title) msgid "Xen terminology" msgstr "" #: ./doc/config-reference/compute/section_introduction-to-xen.xml16(para) msgid "" "Xen. A hypervisor that provides the " "fundamental isolation between virtual machines. Xen is open source (GPLv2) " "and is managed by Xen.org, an cross-industry organization." msgstr "" #: ./doc/config-reference/compute/section_introduction-to-xen.xml20(para) msgid "" "Xen is a component of many different products and projects. The hypervisor " "itself is very similar across all these projects, but the way that it is " "managed can be different, which can cause confusion if you're not clear " "which tool stack you are using. Make sure you know what tool stack you want " "before you get started." msgstr "" #: ./doc/config-reference/compute/section_introduction-to-xen.xml26(para) msgid "" "Xen Cloud Platform (XCP). An open source " "(GPLv2) tool stack for Xen. It is designed specifically as a platform for " "enterprise and cloud computing, and is well integrated with OpenStack. XCP " "is available both as a binary distribution, installed from an iso, and from " "Linux distributions, such as xcp-xapi in Ubuntu. The current versions of XCP available" " in Linux distributions do not yet include all the features available in the" " binary distribution of XCP." msgstr "" #: ./doc/config-reference/compute/section_introduction-to-xen.xml38(para) msgid "" "Citrix XenServer. A commercial product. " "It is based on XCP, and exposes the same tool stack and management API. As " "an analogy, think of XenServer being based on XCP in the way that Red Hat " "Enterprise Linux is based on Fedora. XenServer has a free version (which is " "very similar to XCP) and paid-for versions with additional features enabled." " Citrix provides support for XenServer, but as of July 2012, they do not " "provide any support for XCP. For a comparison between these products see the" " XCP " "Feature Matrix." msgstr "" #: ./doc/config-reference/compute/section_introduction-to-xen.xml50(para) msgid "" "Both XenServer and XCP include Xen, Linux, and the primary control daemon " "known as xapi." msgstr "" #: ./doc/config-reference/compute/section_introduction-to-xen.xml53(para) msgid "" "The API shared between XCP and XenServer is called XenAPI. OpenStack usually refers to XenAPI, to " "indicate that the integration works equally well on XCP and XenServer. " "Sometimes, a careless person will refer to XenServer specifically, but you " "can be reasonably confident that anything that works on XenServer will also " "work on the latest version of XCP. Read the " " XenAPI Object Model Overview for definitions of XenAPI specific " "terms such as SR, VDI, VIF and PIF." msgstr "" #: ./doc/config-reference/compute/section_introduction-to-xen.xml66(title) msgid "Privileged and unprivileged domains" msgstr "" #: ./doc/config-reference/compute/section_introduction-to-xen.xml67(para) msgid "" "A Xen host runs a number of virtual machines, VMs, or domains (the terms are" " synonymous on Xen). One of these is in charge of running the rest of the " "system, and is known as \"domain 0,\" or \"dom0.\" It is the first domain to" " boot after Xen, and owns the storage and networking hardware, the device " "drivers, and the primary control software. Any other VM is unprivileged, and" " are known as a \"domU\" or \"guest\". All customer VMs are unprivileged of " "course, but you should note that on Xen the OpenStack control software " "(nova-compute) also runs in a " "domU. This gives a level of security isolation between the privileged system" " software and the OpenStack software (much of which is customer-facing). " "This architecture is described in more detail later." msgstr "" #: ./doc/config-reference/compute/section_introduction-to-xen.xml83(para) msgid "" "There is an ongoing project to split domain 0 into multiple privileged " "domains known as driver domains and " "stub domains. This would give even better" " separation between critical components. This technology is what powers " "Citrix XenClient RT, and is likely to be added into XCP in the next few " "years. However, the current architecture just has three levels of " "separation: dom0, the OpenStack domU, and the completely unprivileged " "customer VMs." msgstr "" #: ./doc/config-reference/compute/section_introduction-to-xen.xml96(title) msgid "Paravirtualized versus hardware virtualized domains" msgstr "" #: ./doc/config-reference/compute/section_introduction-to-xen.xml98(para) msgid "" "A Xen virtual machine can be paravirtualized " "(PV) or hardware virtualized " "(HVM). This refers to the interaction between Xen, domain 0, and " "the guest VM's kernel. PV guests are aware of the fact that they are " "virtualized and will co-operate with Xen and domain 0; this gives them " "better performance characteristics. HVM guests are not aware of their " "environment, and the hardware has to pretend that they are running on an " "unvirtualized machine. HVM guests do not need to modify the guest operating " "system, which is essential when running Windows." msgstr "" #: ./doc/config-reference/compute/section_introduction-to-xen.xml111(para) msgid "" "In OpenStack, customer VMs may run in either PV or HVM mode. However, the " "OpenStack domU (that's the one running nova-" "compute) must be running in " "PV mode." msgstr "" #: ./doc/config-reference/compute/section_introduction-to-xen.xml119(title) msgid "XenAPI Deployment Architecture" msgstr "" #: ./doc/config-reference/compute/section_introduction-to-xen.xml121(para) msgid "" "When you deploy OpenStack on XCP or XenServer, you get something similar to " "this: " msgstr "" #: ./doc/config-reference/compute/section_introduction-to-xen.xml133(para) msgid "The hypervisor: Xen" msgstr "" #: ./doc/config-reference/compute/section_introduction-to-xen.xml136(para) msgid "" "Domain 0: runs xapi and some small pieces from OpenStack (some xapi plug-ins" " and network isolation rules). The majority of this is provided by XenServer" " or XCP (or yourself using Kronos)." msgstr "" #: ./doc/config-reference/compute/section_introduction-to-xen.xml143(para) msgid "" "OpenStack VM: The nova-compute " "code runs in a paravirtualized virtual machine, running on the host under " "management. Each host runs a local instance of nova-compute. It will often also be running nova-network " "(depending on your network mode). In this case, nova-network is managing the" " addresses given to the tenant VMs through DHCP." msgstr "" #: ./doc/config-reference/compute/section_introduction-to-xen.xml155(para) msgid "" "Nova uses the XenAPI Python library to talk to xapi, and it uses the " "Management Network to reach from the domU to dom0 without leaving the host." msgstr "" #: ./doc/config-reference/compute/section_introduction-to-xen.xml131(para) msgid "Key things to note: " msgstr "" #: ./doc/config-reference/compute/section_introduction-to-xen.xml163(para) msgid "The above diagram assumes FlatDHCP networking (the DevStack default)." msgstr "" #: ./doc/config-reference/compute/section_introduction-to-xen.xml169(para) msgid "" "Management network - RabbitMQ, MySQL, etc. Please note that the VM images " "are downloaded by the XenAPI plug-ins, so make sure that the images can be " "downloaded through the management network. It usually means binding those " "services to the management interface." msgstr "" #: ./doc/config-reference/compute/section_introduction-to-xen.xml179(para) msgid "" "Tenant network - controlled by nova-network. The parameters of this network " "depend on the networking model selected (Flat, Flat DHCP, VLAN)." msgstr "" #: ./doc/config-reference/compute/section_introduction-to-xen.xml186(para) msgid "Public network - floating IPs, public API endpoints." msgstr "" #: ./doc/config-reference/compute/section_introduction-to-xen.xml167(para) msgid "There are three main OpenStack Networks:" msgstr "" #: ./doc/config-reference/compute/section_introduction-to-xen.xml192(para) msgid "" "The networks shown here must be connected to the corresponding physical " "networks within the data center. In the simplest case, three individual " "physical network cards could be used. It is also possible to use VLANs to " "separate these networks. Please note, that the selected configuration must " "be in line with the networking model selected for the cloud. (In case of " "VLAN networking, the physical channels have to be able to forward the tagged" " traffic.)" msgstr "" #: ./doc/config-reference/compute/section_introduction-to-xen.xml161(para) msgid "Some notes on the networking: " msgstr "" #: ./doc/config-reference/compute/section_introduction-to-xen.xml207(title) msgid "XenAPI pools" msgstr "" #: ./doc/config-reference/compute/section_introduction-to-xen.xml208(para) msgid "" "The host-aggregates feature enables you to create pools of XenServer hosts " "to enable live migration when using shared storage. However, you cannot " "configure shared storage." msgstr "" #: ./doc/config-reference/compute/section_introduction-to-xen.xml214(title) msgid "Further reading" msgstr "" #: ./doc/config-reference/compute/section_introduction-to-xen.xml218(para) msgid "" "Citrix XenServer official documentation: " "http://docs.vmd.citrix.com/XenServer." msgstr "" #: ./doc/config-reference/compute/section_introduction-to-xen.xml225(para) msgid "" "What is Xen? by Xen.org: " "http://xen.org/files/Marketing/WhatisXen.pdf." msgstr "" #: ./doc/config-reference/compute/section_introduction-to-xen.xml231(para) msgid "" "Xen Hypervisor project: " "http://xen.org/products/xenhyp.html." msgstr "" #: ./doc/config-reference/compute/section_introduction-to-xen.xml237(para) msgid "" "XCP project: " "http://xen.org/products/cloudxen.html." msgstr "" #: ./doc/config-reference/compute/section_introduction-to-xen.xml243(para) msgid "" "Further XenServer and OpenStack information: " "http://wiki.openstack.org/XenServer." msgstr "" #: ./doc/config-reference/compute/section_introduction-to-xen.xml215(para) msgid "" "Here are some of the resources available to learn more about Xen: " "" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_baremetal.xml6(title) msgid "Bare metal driver" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_baremetal.xml7(para) msgid "" "The baremetal driver is a hypervisor driver for OpenStack Nova Compute. " "Within the OpenStack framework, it has the same role as the drivers for " "other hypervisors (libvirt, xen, etc), and yet it is presently unique in " "that the hardware is not virtualized - there is no hypervisor between the " "tenants and the physical hardware. It exposes hardware through the OpenStack" " APIs, using pluggable sub-drivers to deliver machine imaging (PXE) and " "power control (IPMI). With this, provisioning and management of physical " "hardware is accomplished by using common cloud APIs and tools, such as " "OpenStack Orchestration or salt-cloud. However, due to this unique " "situation, using the baremetal driver requires some additional preparation " "of its environment, the details of which are beyond the scope of this guide." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_baremetal.xml20(para) msgid "" "Some OpenStack Compute features are not implemented by the baremetal " "hypervisor driver. See the hypervisor " "support matrix for details." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_baremetal.xml24(para) msgid "" "For the Baremetal driver to be loaded and function properly, ensure that the" " following options are set in /etc/nova/nova.conf on " "your nova-compute hosts." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_baremetal.xml34(para) msgid "" "Many configuration options are specific to the Baremetal driver. Also, some " "additional steps are required, such as building the baremetal deploy " "ramdisk. See the main wiki page for" " details and implementation suggestions." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-backing-storage.xml10(title) msgid "Configure Compute backing storage" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-backing-storage.xml11(para) msgid "" "Backing Storage is the storage used to provide the expanded operating system" " image, and any ephemeral storage. Inside the virtual machine, this is " "normally presented as two virtual hard disks (for example, /dev/vda and " "/dev/vdb respectively). However, inside OpenStack, this can be derived from " "one of three methods: LVM, QCOW or RAW, chosen using the " "libvirt_images_type option in " "nova.conf on the compute node." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-backing-storage.xml19(para) msgid "" "QCOW is the default backing store. It uses a copy-on-write philosophy to " "delay allocation of storage until it is actually needed. This means that the" " space required for the backing of an image can be significantly less on the" " real disk than what seems available in the virtual machine operating " "system." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-backing-storage.xml24(para) msgid "" "RAW creates files without any sort of file formatting, effectively creating " "files with the plain binary one would normally see on a real disk. This can " "increase performance, but means that the entire size of the virtual disk is " "reserved on the physical disk." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-backing-storage.xml29(para) msgid "" "Local LVM" " volumes can also be used. Set " "libvirt_images_volume_group=nova_local where " "nova_local is the name of the LVM group you have created." msgstr "" #: ./doc/config-reference/compute/section_compute-hypervisors.xml10(title) msgid "Hypervisors" msgstr "" #: ./doc/config-reference/compute/section_compute-hypervisors.xml11(para) msgid "" "OpenStack Compute supports many hypervisors, which might make it difficult " "for you to choose one. Most installations use only one hypervisor. However " "you can use and to schedule to different hypervisors " "within the same installation. The following links help you choose a " "hypervisor. See http://wiki.openstack.org/HypervisorSupportMatrix" " for a detailed list of features and support across the hypervisors." msgstr "" #: ./doc/config-reference/compute/section_compute-hypervisors.xml22(para) msgid "The following hypervisors are supported:" msgstr "" #: ./doc/config-reference/compute/section_compute-hypervisors.xml25(para) msgid "" "KVM - Kernel-" "based Virtual Machine. The virtual disk formats that it supports is " "inherited from QEMU since it uses a modified QEMU program to launch the " "virtual machine. The supported formats include raw images, the qcow2, and " "VMware formats." msgstr "" #: ./doc/config-reference/compute/section_compute-hypervisors.xml34(para) msgid "" "LXC - Linux Containers " "(through libvirt), use to run Linux-based virtual machines." msgstr "" #: ./doc/config-reference/compute/section_compute-hypervisors.xml39(para) msgid "" "QEMU - Quick EMUlator, " "generally only used for development purposes." msgstr "" #: ./doc/config-reference/compute/section_compute-hypervisors.xml44(para) msgid "" "UML - User " "Mode Linux, generally only used for development purposes." msgstr "" #: ./doc/config-reference/compute/section_compute-hypervisors.xml50(para) msgid "" "VMWare vSphere 4.1 update 1 and newer, runs" " VMWare-based Linux and Windows images through a connection with a vCenter " "server or directly with an ESXi host." msgstr "" #: ./doc/config-reference/compute/section_compute-hypervisors.xml58(para) msgid "" "Xen - XenServer, Xen Cloud Platform" " (XCP), use to run Linux or Windows virtual machines. You must install the " "nova-compute service in a para-" "virtualized VM." msgstr "" #: ./doc/config-reference/compute/section_compute-hypervisors.xml66(para) msgid "" " Hyper-V - Server " "virtualization with Microsoft's Hyper-V, use to run Windows, Linux, and " "FreeBSD virtual machines. Runs nova-" "compute natively on the Windows virtualization platform." msgstr "" #: ./doc/config-reference/compute/section_compute-hypervisors.xml75(para) msgid "" " Bare Metal " "- Not a hypervisor in the traditional sense, this driver provisions physical" " hardware through pluggable sub-drivers (for example, PXE for image " "deployment, and IPMI for power management)." msgstr "" #: ./doc/config-reference/compute/section_compute-hypervisors.xml83(para) msgid "" "Dockeris an open-source engine " "which automates the deployment of >applications as highly portable, self-" "sufficient containers which are >independent of hardware, language, " "framework, packaging system and hosting >provider." msgstr "" #: ./doc/config-reference/compute/section_compute-hypervisors.xml92(title) msgid "Hypervisor configuration basics" msgstr "" #: ./doc/config-reference/compute/section_compute-hypervisors.xml93(para) msgid "" "The node where the nova-compute " "service is installed and running is the machine that runs all the virtual " "machines, referred to as the compute node in this guide." msgstr "" #: ./doc/config-reference/compute/section_compute-hypervisors.xml97(para) msgid "" "By default, the selected hypervisor is KVM. To change to another hypervisor," " change the libvirt_type option in " "nova.conf and restart the nova-compute service." msgstr "" #: ./doc/config-reference/compute/section_compute-hypervisors.xml103(para) msgid "" "Here are the general nova.conf options that are used to" " configure the compute node's hypervisor. Specific options for particular " "hypervisors can be found in following sections." msgstr "" #. When image changes, this message will be marked fuzzy or untranslated for #. you. #. It doesn't matter what you translate it to: it's not used at all. #: ./doc/config-reference/compute/section_compute-config-samples.xml44(None) msgid "" "@@image: '../../common/figures/SCH_5004_V00_NUAC-" "Network_mode_KVM_Flat_OpenStack.png'; md5=1e883ef27e5912b5c516d153b8844a28" msgstr "" #. When image changes, this message will be marked fuzzy or untranslated for #. you. #. It doesn't matter what you translate it to: it's not used at all. #: ./doc/config-reference/compute/section_compute-config-samples.xml83(None) msgid "" "@@image: '../../common/figures/SCH_5005_V00_NUAC-" "Network_mode_XEN_Flat_OpenStack.png'; md5=3b151435a0fda3702d4fac5a964fac83" msgstr "" #: ./doc/config-reference/compute/section_compute-config-samples.xml9(title) msgid "Example nova.conf configuration files" msgstr "" #: ./doc/config-reference/compute/section_compute-config-samples.xml11(para) msgid "" "The following sections describe the configuration options in the " "nova.conf file. You must copy the " "nova.conf file to each compute node. The sample " "nova.conf files show examples of specific " "configurations." msgstr "" #: ./doc/config-reference/compute/section_compute-config-samples.xml17(title) msgid "Small, private cloud" msgstr "" #: ./doc/config-reference/compute/section_compute-config-samples.xml18(para) msgid "" "This example nova.conf file configures a small private " "cloud with cloud controller services, database server, and messaging server " "on the same server. In this case, CONTROLLER_IP represents the IP address of" " a central server, BRIDGE_INTERFACE represents the bridge such as br100, the" " NETWORK_INTERFACE represents an interface to your VLAN setup, and passwords" " are represented as DB_PASSWORD_COMPUTE for your Compute (nova) database " "password, and RABBIT PASSWORD represents the password to your message queue " "installation." msgstr "" #: ./doc/config-reference/compute/section_compute-config-samples.xml31(title) #: ./doc/config-reference/compute/section_compute-config-samples.xml38(title) #: ./doc/config-reference/compute/section_compute-config-samples.xml77(title) msgid "KVM, Flat, MySQL, and Glance, OpenStack or EC2 API" msgstr "" #: ./doc/config-reference/compute/section_compute-config-samples.xml33(para) msgid "" "This example nova.conf file, from an internal Rackspace" " test system, is used for demonstrations." msgstr "" #: ./doc/config-reference/compute/section_compute-config-samples.xml50(title) msgid "XenServer, Flat networking, MySQL, and Glance, OpenStack API" msgstr "" #: ./doc/config-reference/compute/section_compute-config-samples.xml52(para) msgid "" "This example nova.conf file is from an internal " "Rackspace test system." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_lxc.xml6(title) msgid "LXC (Linux containers)" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_lxc.xml7(para) msgid "" "LXC (also known as Linux containers) is a virtualization technology that " "works at the operating system level. This is different from hardware " "virtualization, the approach used by other hypervisors such as KVM, Xen, and" " VMWare. LXC (as currently implemented using libvirt in the nova project) is" " not a secure virtualization technology for multi-tenant environments " "(specifically, containers may affect resource quotas for other containers " "hosted on the same machine). Additional containment technologies, such as " "AppArmor, may be used to provide better isolation between containers, " "although this is not the case by default. For all these reasons, the choice " "of this virtualization technology is not recommended in production." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_lxc.xml19(para) msgid "" "If your compute hosts do not have hardware support for virtualization, LXC " "will likely provide better performance than QEMU. In addition, if your " "guests must access specialized hardware, such as GPUs, this might be easier " "to achieve with LXC than other hypervisors." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_lxc.xml22(para) msgid "" "Some OpenStack Compute features might be missing when running with LXC as " "the hypervisor. See the hypervisor " "support matrix for details." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_lxc.xml25(para) msgid "" "To enable LXC, ensure the following options are set in " "/etc/nova/nova.conf on all hosts running the " "nova-compute " "service." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_lxc.xml30(para) msgid "" "On Ubuntu 12.04, enable LXC support in OpenStack by installing the nova-compute-lxc package." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-db.xml9(title) msgid "Database configuration" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-db.xml10(para) msgid "" "You can configure OpenStack Compute to use any SQLAlchemy-compatible " "database. The database name is nova. The nova-conductor service is the only service " "that writes to the database. The other Compute services access the database " "through the nova-conductor " "service." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-db.xml17(para) msgid "" "To ensure that the database schema is current, run the following command:" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-db.xml19(para) msgid "" "If nova-conductor is not used, " "entries to the database are mostly written by the nova-scheduler service, although all services" " must be able to update entries in the database." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-db.xml24(para) msgid "" "In either case, use these settings to configure the connection string for " "the nova database." msgstr "" #: ./doc/config-reference/compute/section_compute-options-reference.xml6(title) #: ./doc/config-reference/block-storage/drivers/xenapi-nfs.xml86(title) #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml309(caption) msgid "Configuration options" msgstr "" #: ./doc/config-reference/compute/section_compute-options-reference.xml7(para) msgid "" "For a complete list of all available configuration options for each " "OpenStack Compute service, run bin/nova-<servicename> --help." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-ipv6.xml9(title) msgid "Configure Compute to use IPv6 addresses" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-ipv6.xml10(para) msgid "" "You can configure Compute to use both IPv4 and IPv6 addresses for " "communication by putting it into a IPv4/IPv6 dual stack mode. In IPv4/IPv6 " "dual stack mode, instances can acquire their IPv6 global unicast address by " "stateless address auto configuration mechanism [RFC 4862/2462]. IPv4/IPv6 " "dual stack mode works with VlanManager and " "FlatDHCPManager networking modes. In " "VlanManager, different 64bit global routing prefix is " "used for each project. In FlatDHCPManager, one 64bit " "global routing prefix is used for all instances." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-ipv6.xml18(para) msgid "" "This configuration has been tested with VM images that have IPv6 stateless " "address auto configuration capability (must use EUI-64 address for stateless" " address auto configuration), a requirement for any VM you want to run with " "an IPv6 address. Each node that executes a nova-* service" " must have python-netaddr and radvd " "installed." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-ipv6.xml24(para) msgid "On all nova-nodes, install python-netaddr:" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-ipv6.xml26(para) msgid "" "On all nova-network nodes install " "radvd and configure IPv6 networking:" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-ipv6.xml31(para) msgid "" "Edit the nova.conf file on all nodes to set the " "use_ipv6 configuration option to True. Restart all nova- services." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-ipv6.xml34(para) msgid "" "When using the command you can add a fixed range for IPv6 " "addresses. You must specify public or private after the create parameter." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-ipv6.xml36(replaceable) msgid "fixed_range_v4" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-ipv6.xml36(replaceable) msgid "vlan_id" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-ipv6.xml36(replaceable) msgid "vpn_start" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-ipv6.xml36(replaceable) msgid "fixed_range_v6" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-ipv6.xml37(para) msgid "" "You can set IPv6 global routing prefix by using the " "--fixed_range_v6 parameter. The default is: " "fd00::/48. When you use " "FlatDHCPManager, the command uses the original value of " "--fixed_range_v6. When you use " "VlanManager, the command creates prefixes of subnet by " "incrementing subnet id. Guest VMs uses this prefix for generating their IPv6" " global unicast address." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-ipv6.xml43(para) msgid "Here is a usage example for VlanManager:" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-ipv6.xml45(para) msgid "Here is a usage example for FlatDHCPManager:" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-console.xml7(title) msgid "Configure remote console access" msgstr "" #. #: ./doc/config-reference/compute/section_compute-configure-console.xml9(para) msgid "" "OpenStack has two main methods for providing a remote console or remote " "desktop access to guest Virtual Machines. They are VNC, and SPICE HTML5 and " "can be used either through the OpenStack dashboard and the command line. " "Best practice is to select one or the other to run." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_kvm.xml6(para) msgid "KVM is configured as the default hypervisor for Compute." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_kvm.xml8(para) msgid "" "This document contains several sections about hypervisor selection. If you " "are reading this document linearly, you do not want to load the KVM module " "before you install nova-compute. " "The nova-compute service depends " "on qemu-kvm, which installs /lib/udev/rules.d/45-qemu-" "kvm.rules, which sets the correct permissions on the /dev/kvm " "device node." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_kvm.xml15(para) msgid "" "To enable KVM explicitly, add the following configuration options to the " "/etc/nova/nova.conf file:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_kvm.xml19(para) msgid "" "The KVM hypervisor supports the following virtual machine image formats:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_kvm.xml22(para) #: ./doc/config-reference/compute/section_hypervisor_qemu.xml43(para) msgid "Raw" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_kvm.xml25(para) #: ./doc/config-reference/compute/section_hypervisor_qemu.xml46(para) msgid "QEMU Copy-on-write (qcow2)" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_kvm.xml28(para) msgid "QED Qemu Enhanced Disk" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_kvm.xml31(para) #: ./doc/config-reference/compute/section_hypervisor_qemu.xml49(para) msgid "VMWare virtual machine disk format (vmdk)" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_kvm.xml34(para) msgid "" "This section describes how to enable KVM on your system. For more " "information, see the following distribution-specific documentation:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_kvm.xml38(para) msgid "" "Fedora:" " Getting started with virtualization from the Fedora project wiki." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_kvm.xml44(para) msgid "" "Ubuntu: " "KVM/Installation from the Community Ubuntu documentation." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_kvm.xml48(para) msgid "" "Debian: " "Virtualization with KVM from the Debian handbook." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_kvm.xml53(para) msgid "" "RHEL:" " Installing virtualization packages on an existing Red Hat Enterprise Linux " "system from the Red Hat Enterprise Linux Virtualization " "Host Configuration and Guest Installation Guide." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_kvm.xml60(para) msgid "" "openSUSE: Installing " "KVM from the openSUSE Virtualization with KVM manual." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_kvm.xml66(para) msgid "" "SLES: " "Installing KVM from the SUSE Linux Enterprise Server Virtualization " "with KVM manual." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_kvm.xml76(title) msgid "Specify the CPU model of KVM guests" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_kvm.xml77(para) msgid "" "The Compute service enables you to control the guest CPU model that is " "exposed to KVM virtual machines. Use cases include:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_kvm.xml81(para) msgid "" "To maximize performance of virtual machines by exposing new host CPU " "features to the guest" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_kvm.xml85(para) msgid "" "To ensure a consistent default CPU across all machines, removing reliance of" " variable QEMU defaults" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_kvm.xml89(para) msgid "" "In libvirt, the CPU is specified by providing a base CPU model name (which " "is a shorthand for a set of feature flags), a set of additional feature " "flags, and the topology (sockets/cores/threads). The libvirt KVM driver " "provides a number of standard CPU model names. These models are defined in " "the /usr/share/libvirt/cpu_map.xml file. Check this " "file to determine which models are supported by your local installation." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_kvm.xml95(para) msgid "" "Two Compute configuration options define which type of CPU model is exposed " "to the hypervisor when using KVM: libvirt_cpu_mode and " "libvirt_cpu_model." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_kvm.xml98(para) msgid "" "The libvirt_cpu_mode option can take one of the following" " values: none, host-passthrough, " "host-model, and custom." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_kvm.xml102(title) msgid "Host model (default for KVM & QEMU)" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_kvm.xml103(para) msgid "" "If your nova.conf file contains " "libvirt_cpu_mode=host-model, libvirt identifies the CPU " "model in /usr/share/libvirt/cpu_map.xml file that most " "closely matches the host, and requests additional CPU flags to complete the " "match. This configuration provides the maximum functionality and performance" " and maintains good reliability and compatibility if the guest is migrated " "to another host with slightly different host CPUs." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_kvm.xml112(title) msgid "Host pass through" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_kvm.xml113(para) msgid "" "If your nova.conf file contains " "libvirt_cpu_mode=host-passthrough, libvirt tells KVM to " "pass through the host CPU with no modifications. The difference to host-" "model, instead of just matching feature flags, every last detail of the host" " CPU is matched. This gives absolutely best performance, and can be " "important to some apps which check low level CPU details, but it comes at a " "cost with respect to migration: the guest can only be migrated to an exactly" " matching host CPU." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_kvm.xml122(title) msgid "Custom" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_kvm.xml123(para) msgid "" "If your nova.conf file contains " "libvirt_cpu_mode=custom, you can explicitly specify one " "of the supported named model using the libvirt_cpu_model configuration " "option. For example, to configure the KVM guests to expose Nehalem CPUs, " "your nova.conf file should contain:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_kvm.xml132(title) msgid "" "None (default for all libvirt-driven hypervisors other than KVM & QEMU)" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_kvm.xml134(para) msgid "" "If your nova.conf file contains " "libvirt_cpu_mode=none, libvirt does not specify a CPU " "model. Instead, the hypervisor chooses the default model. This setting is " "equivalent to the Compute service behavior prior to the Folsom release." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_kvm.xml141(title) msgid "Guest agent support" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_kvm.xml142(para) msgid "" "With the Havana release, support for guest agents was added, allowing " "optional access between compute nods and guests through a socket, using the " "qmp protocol." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_kvm.xml144(para) msgid "" "To enable this feature, you must set " "hw_qemu_guest_agent=yes as a metadata parameter on the " "image you wish to use to create guest-agent-capable instances from. You can " "explicitly disable the feature by setting " "hw_qemu_guest_agent=no in the image metadata." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_kvm.xml150(title) msgid "KVM performance tweaks" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_kvm.xml151(para) msgid "" "The VHostNet " "kernel module improves network performance. To load the kernel module, run " "the following command as root:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_kvm.xml157(title) msgid "Troubleshoot KVM" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_kvm.xml158(para) msgid "" "Trying to launch a new virtual machine instance fails with the " "ERRORstate, and the following error appears in the " "/var/log/nova/nova-compute.log file:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_kvm.xml162(para) msgid "This message indicates that the KVM kernel modules were not loaded." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_kvm.xml163(para) msgid "" "If you cannot start VMs after installation without rebooting, the " "permissions might not be correct. This can happen if you load the KVM module" " before you install nova-compute." " To check whether the group is set to kvm, run:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_kvm.xml168(para) msgid "If it is not set to kvm, run:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_docker.xml6(title) msgid "Docker driver" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_docker.xml7(para) msgid "" "The Docker driver is a hypervisor driver for OpenStack Compute, introduced " "with the Havana release. Docker is an open-source engine which automates the" " deployment of applications as highly portable, self-sufficient containers " "which are independent of hardware, language, framework, packaging system and" " hosting provider. Docker extends LXC with a high level API providing a " "lightweight virtualization solution that runs processes in isolation. It " "provides a way to automate software deployment in a secure and repeatable " "environment. A standard container in Docker contains a software component " "along with all of its dependencies - binaries, libraries, configuration " "files, scripts, virtualenvs, jars, gems and tarballs. Docker can be run on " "any x86_64 Linux kernel that supports cgroups and aufs. Docker is a way of " "managing LXC containers on a single machine. However used behind OpenStack " "Compute makes Docker much more powerful since it is then possible to manage " "several hosts which will then manage hundreds of containers. The current " "Docker project aims for full OpenStack compatibility. Containers don't aim " "to be a replacement for VMs, they are just complementary in the sense that " "they are better for specific use cases. Compute's support for VMs is " "currently advanced thanks to the variety of hypervisors running VMs. However" " it's not the case for containers even though libvirt/LXC is a good starting" " point. Docker aims to go the second level of integration." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_docker.xml27(para) msgid "" "Some OpenStack Compute features are not implemented by the docker driver. " "See the " "hypervisor support matrix for details." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_docker.xml33(para) msgid "" "To enable Docker, ensure the following options are set in " "/etc/nova/nova-compute.conf on all hosts running the " "nova-compute service. " "" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_docker.xml37(para) msgid "" "Glance also needs to be configured to support the Docker container format, " "in /etc/glance-api.conf: " msgstr "" #: ./doc/config-reference/compute/section_compute-configure-service-groups.xml15(title) msgid "Configuring Compute Service Groups" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-service-groups.xml16(para) msgid "" "To effectively manage and utilize compute nodes, the Compute service must " "know their statuses. For example, when a user launches a new VM, the Compute" " scheduler should send the request to a live node (with enough capacity too," " of course). From the Grizzly release and later, the Compute service queries" " the ServiceGroup API to get the node liveness information." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-service-groups.xml21(para) msgid "" "When a compute worker (running the nova-" "compute daemon) starts, it calls the join API to join the " "compute group, so that every service that is interested in the information " "(for example, the scheduler) can query the group membership or the status of" " a particular node. Internally, the ServiceGroup client driver automatically" " updates the compute worker status." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-service-groups.xml27(para) msgid "" "The following drivers are implemented: database and ZooKeeper. Further " "drivers are in review or development, such as memcache." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-service-groups.xml31(title) msgid "Database ServiceGroup driver" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-service-groups.xml32(para) msgid "" "Compute uses the database driver, which is the default driver, to track node" " liveness. In a compute worker, this driver periodically sends a " " command to the database, saying I'm OK with " "a timestamp. A pre-defined timeout (service_down_time) " "determines if a node is dead." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-service-groups.xml38(para) msgid "" "The driver has limitations, which may or may not be an issue for you, " "depending on your setup. The more compute worker nodes that you have, the " "more pressure you put on the database. By default, the timeout is 60 seconds" " so it might take some time to detect node failures. You could reduce the " "timeout value, but you must also make the DB update more frequently, which " "again increases the DB workload." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-service-groups.xml44(para) msgid "" "Fundamentally, the data that describes whether the node is alive is " "\"transient\" — After a few seconds, this data is obsolete. Other data in " "the database is persistent, such as the entries that describe who owns which" " VMs. However, because this data is stored in the same database, is treated " "the same way. The ServiceGroup abstraction aims to treat them separately." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-service-groups.xml53(title) msgid "ZooKeeper ServiceGroup driver" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-service-groups.xml54(para) msgid "" "The ZooKeeper ServiceGroup driver works by using ZooKeeper ephemeral nodes. " "ZooKeeper, in contrast to databases, is a distributed system. Its load is " "divided among several servers. At a compute worker node, after establishing " "a ZooKeeper session, it creates an ephemeral znode in the group directory. " "Ephemeral znodes have the same lifespan as the session. If the worker node " "or the nova-compute daemon " "crashes, or a network partition is in place between the worker and the " "ZooKeeper server quorums, the ephemeral znodes are removed automatically. " "The driver gets the group membership by running the command" " in the group directory." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-service-groups.xml64(para) msgid "" "To use the ZooKeeper driver, you must install ZooKeeper servers and client " "libraries. Setting up ZooKeeper servers is outside the scope of this " "article. For the rest of the article, assume these servers are installed, " "and their addresses and ports are 192.168.2.1:2181, " "192.168.2.2:2181, 192.168.2.3:2181." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-service-groups.xml71(para) msgid "" "To use ZooKeeper, you must install client-side Python libraries on every " "nova node: python-zookeeper – the official Zookeeper " "Python binding and evzookeeper – the library to make the " "binding work with the eventlet threading model." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-service-groups.xml77(para) msgid "" "The relevant configuration snippet in the " "/etc/nova/nova.conf file on every node is:" msgstr "" #: ./doc/config-reference/compute/section_compute-conductor.xml7(title) msgid "Conductor" msgstr "" #: ./doc/config-reference/compute/section_compute-conductor.xml8(para) msgid "" "The nova-conductor service " "enables OpenStack to function without compute nodes accessing the database. " "Conceptually, it implements a new layer on top of nova-compute. It should not be deployed on " "compute nodes, or else the security benefits of removing database access " "from nova-compute are negated. " "Just like other nova services such as nova-" "api or nova-scheduler, it can be scaled horizontally. You can " "run multiple instances of nova-" "conductor on different machines as needed for scaling purposes." msgstr "" #: ./doc/config-reference/compute/section_compute-conductor.xml21(para) msgid "" "In the Grizzly release, the methods exposed by nova-conductor are relatively simple methods used by " "nova-compute to offload its " "database operations. Places where nova-" "compute previously performed database access are now talking to" " nova-conductor. However, we have" " plans in the medium to long term to move more and more of what is currently" " in nova-compute up to the " "nova-conductor layer. The compute" " service will start to look like a less intelligent slave service to " "nova-conductor. The conductor " "service will implement long running complex operations, ensuring forward " "progress and graceful error handling. This will be especially beneficial for" " operations that cross multiple compute nodes, such as migrations or " "resizes." msgstr "" #: ./doc/config-reference/compute/section_nova-conf.xml6(title) msgid "Overview of nova.conf" msgstr "" #: ./doc/config-reference/compute/section_nova-conf.xml7(para) msgid "" "The nova.conf configuration file is an INI file format file " "that specifies options as key=value pairs, which are " "grouped into sections. The DEFAULT section contains most " "of the configuration options. For example:" msgstr "" #: ./doc/config-reference/compute/section_nova-conf.xml19(para) msgid "" "You can use a particular configuration option file by using the " "option (nova.conf) parameter when " "you run one of the nova-* services. This parameter " "inserts configuration option definitions from the specified configuration " "file name, which might be useful for debugging or performance tuning." msgstr "" #: ./doc/config-reference/compute/section_nova-conf.xml25(para) msgid "" "To place comments in the nova.conf file, start a new " "line that begins with the pound (#) character. For a list" " of configuration options, see the tables in this guide." msgstr "" #: ./doc/config-reference/compute/section_nova-conf.xml29(para) msgid "" "To learn more about the nova.conf configuration file, " "review these general purpose configuration options." msgstr "" #: ./doc/config-reference/compute/section_nova-conf.xml34(title) msgid "Types of configuration options" msgstr "" #: ./doc/config-reference/compute/section_nova-conf.xml35(para) msgid "" "Each configuration option has an associated data type. The supported data " "types for configuration options are:" msgstr "" #: ./doc/config-reference/compute/section_nova-conf.xml40(term) msgid "BoolOpt" msgstr "" #: ./doc/config-reference/compute/section_nova-conf.xml42(para) msgid "" "Boolean option. Value must be either true or " "false . Example:" msgstr "" #: ./doc/config-reference/compute/section_nova-conf.xml49(term) msgid "StrOpt" msgstr "" #: ./doc/config-reference/compute/section_nova-conf.xml51(para) msgid "String option. Value is an arbitrary string. Example:" msgstr "" #: ./doc/config-reference/compute/section_nova-conf.xml56(term) msgid "IntOption" msgstr "" #: ./doc/config-reference/compute/section_nova-conf.xml58(para) msgid "Integer option. Value must be an integer. Example: " msgstr "" #: ./doc/config-reference/compute/section_nova-conf.xml64(term) msgid "MultiStrOpt" msgstr "" #: ./doc/config-reference/compute/section_nova-conf.xml66(para) msgid "" "String option. Same as StrOpt, except that it can be declared multiple times" " to indicate multiple values. Example:" msgstr "" #: ./doc/config-reference/compute/section_nova-conf.xml74(term) msgid "ListOpt" msgstr "" #: ./doc/config-reference/compute/section_nova-conf.xml76(para) msgid "" "List option. Value is a list of arbitrary strings separated by commas. " "Example:" msgstr "" #: ./doc/config-reference/compute/section_nova-conf.xml82(term) msgid "FloatOpt" msgstr "" #: ./doc/config-reference/compute/section_nova-conf.xml84(para) msgid "Floating-point option. Value must be a floating-point number. Example:" msgstr "" #: ./doc/config-reference/compute/section_nova-conf.xml91(para) msgid "Do not specify quotes around Nova options." msgstr "" #: ./doc/config-reference/compute/section_nova-conf.xml95(title) msgid "Sections" msgstr "" #: ./doc/config-reference/compute/section_nova-conf.xml99(literal) msgid "[DEFAULT]" msgstr "" #: ./doc/config-reference/compute/section_nova-conf.xml101(para) msgid "" "Contains most configuration options. If the documentation for a " "configuration option does not specify its section, assume that it appears in" " this section." msgstr "" #: ./doc/config-reference/compute/section_nova-conf.xml109(literal) msgid "[cells]" msgstr "" #: ./doc/config-reference/compute/section_nova-conf.xml111(para) msgid "" "Configures cells functionality. For details, see the Cells section ()." msgstr "" #: ./doc/config-reference/compute/section_nova-conf.xml119(literal) msgid "[baremetal]" msgstr "" #: ./doc/config-reference/compute/section_nova-conf.xml121(para) msgid "Configures the baremetal hypervisor driver." msgstr "" #: ./doc/config-reference/compute/section_nova-conf.xml126(literal) msgid "[conductor]" msgstr "" #: ./doc/config-reference/compute/section_nova-conf.xml128(para) msgid "" "Configures the nova-conductor " "service." msgstr "" #: ./doc/config-reference/compute/section_nova-conf.xml135(literal) msgid "[trusted_computing]" msgstr "" #: ./doc/config-reference/compute/section_nova-conf.xml137(para) msgid "" "Configures the trusted computing pools functionality and how to connect to a" " remote attestation service." msgstr "" #: ./doc/config-reference/compute/section_nova-conf.xml96(para) msgid "" "Configuration options are grouped by section. The Compute configuration file" " supports the following sections:" msgstr "" #: ./doc/config-reference/compute/section_nova-conf.xml146(title) msgid "Variable substitution" msgstr "" #: ./doc/config-reference/compute/section_nova-conf.xml147(para) msgid "" "The configuration file supports variable substitution. After you set a " "configuration option, it can be referenced in later configuration values " "when you precede it with $. This example defines " "my_ip and then uses $my_ip as a " "variable:" msgstr "" #: ./doc/config-reference/compute/section_nova-conf.xml156(para) msgid "" "If a value must contain the $ character, escape it with " "$$. For example, if your LDAP DNS password is " "$xkj432, specify it, as follows:" msgstr "" #: ./doc/config-reference/compute/section_nova-conf.xml161(para) msgid "" "The Compute code uses the Python " "string.Template.safe_substitute() method to implement " "variable substitution. For more details on how variable substitution is " "resolved, see http://docs.python.org/2/library/string.html#template-" "strings and http://www.python.org/dev/peps/pep-0292/." msgstr "" #: ./doc/config-reference/compute/section_nova-conf.xml173(title) msgid "Whitespace" msgstr "" #: ./doc/config-reference/compute/section_nova-conf.xml174(para) msgid "" "To include whitespace in a configuration value, use a quoted string. For " "example:" msgstr "" #: ./doc/config-reference/compute/section_nova-conf.xml179(title) msgid "Define an alternate location for nova.conf" msgstr "" #: ./doc/config-reference/compute/section_nova-conf.xml180(para) msgid "" "All nova-* services and the " " command-line client load the configuration file. To define " "an alternate location for the configuration file, pass the --config-file /path/to/nova.conf " "parameter when you start a nova-*" " service or call a command." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml6(title) msgid "Hyper-V virtualization platform" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml15(emphasis) msgid "Windows Server 2008r2" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml16(para) msgid "" "Both Server and Server Core with the Hyper-V role enabled (Shared Nothing " "Live migration is not supported using 2008r2)" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml20(emphasis) msgid "Windows Server 2012" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml21(para) msgid "Server and Core (with the Hyper-V role enabled), and Hyper-V Server" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml7(para) msgid "" "It is possible to use Hyper-V as a compute node within an OpenStack " "Deployment. The nova-compute " "service runs as \"openstack-compute,\" a 32-bit service directly upon the " "Windows platform with the Hyper-V role enabled. The necessary Python " "components as well as the nova-" "compute service are installed directly onto the Windows " "platform. Windows Clustering Services are not needed for functionality " "within the OpenStack infrastructure. The use of the Windows Server 2012 " "platform is recommend for the best experience and is the platform for active" " development. The following Windows platforms have been tested as compute " "nodes:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml26(title) msgid "Hyper-V configuration" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml27(para) msgid "" "The following sections discuss how to prepare the Windows Hyper-V node for " "operation as an OpenStack Compute node. Unless stated otherwise, any " "configuration information should work for both the Windows 2008r2 and 2012 " "platforms." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml30(emphasis) msgid "Local Storage Considerations" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml31(para) msgid "" "The Hyper-V compute node needs to have ample storage for storing the virtual" " machine images running on the compute nodes. You may use a single volume " "for all, or partition it into an OS volume and VM volume. It is up to the " "individual deploying to decide." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml37(title) msgid "Configure NTP" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml38(para) msgid "" "Network time services must be configured to ensure proper operation of the " "Hyper-V compute node. To set network time on your Hyper-V host you must run " "the following commands:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml52(title) msgid "Configure Hyper-V virtual switching" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml53(para) msgid "" "Information regarding the Hyper-V virtual Switch can be located here: http://technet.microsoft.com/en-" "us/library/hh831823.aspx" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml57(para) msgid "" "To quickly enable an interface to be used as a Virtual Interface the " "following PowerShell may be used:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml67(title) msgid "Enable iSCSI initiator service" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml68(para) msgid "" "To prepare the Hyper-V node to be able to attach to volumes provided by " "cinder you must first make sure the Windows iSCSI initiator service is " "running and started automatically." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml79(title) msgid "Configure shared nothing live migration" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml80(para) msgid "" "Detailed information on the configuration of live migration can be found " "here: http://technet.microsoft.com/en-" "us/library/jj134199.aspx" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml83(para) msgid "The following outlines the steps of shared nothing live migration." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml86(para) msgid "" "The target hosts ensures that live migration is enabled and properly " "configured in Hyper-V." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml90(para) msgid "" "The target hosts checks if the image to be migrated requires a base VHD and " "pulls it from Glance if not already available on the target host." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml94(para) msgid "" "The source hosts ensures that live migration is enabled and properly " "configured in Hyper-V." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml98(para) msgid "The source hosts initiates a Hyper-V live migration." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml101(para) msgid "" "The source hosts communicates to the manager the outcome of the operation." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml109(literal) msgid "instances_shared_storage=False" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml110(para) msgid "" "This needed to support \"shared nothing\" Hyper-V live migrations. It is " "used in nova/compute/manager.py" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml114(literal) msgid "limit_cpu_features=True" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml115(para) msgid "" "This flag is needed to support live migration to hosts with different CPU " "features. This flag is checked during instance creation in order to limit " "the CPU features used by the VM." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml121(literal) msgid "instances_path=DRIVELETTER:\\PATH\\TO\\YOUR\\INSTANCES" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml105(para) msgid "" "The following two configuration options/flags are needed in order to support" " Hyper-V live migration and must be added to your " "nova.conf on the Hyper-V compute node:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml125(para) msgid "Additional Requirements:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml128(para) msgid "Hyper-V 2012 RC or Windows Server 2012 RC with Hyper-V role enabled" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml131(para) msgid "" "A Windows domain controller with the Hyper-V compute nodes as domain members" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml135(para) msgid "" "The instances_path command line option/flag needs to be the same on all " "hosts" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml139(para) msgid "" "The openstack-compute service deployed with the setup must run with domain " "credentials. You can set the service credentials with:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml145(emphasis) msgid "How to setup live migration on Hyper-V" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml146(para) msgid "" "To enable shared nothing live migration run the 3 PowerShell instructions " "below on each Hyper-V host:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml152(replaceable) #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml337(replaceable) #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml338(replaceable) #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml345(replaceable) msgid "IP_ADDRESS" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml158(para) msgid "" "Please replace the IP_ADDRESS with the address of the interface which will " "provide the virtual switching for nova-network." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml161(emphasis) msgid "Additional Reading" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml162(para) msgid "" "Here's an article that clarifies the various live migration options in " "Hyper-V:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml166(link) msgid "" "http://ariessysadmin.blogspot.ro/2012/04/hyper-v-live-migration-of-" "windows.html" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml170(title) msgid "Python Requirements" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml171(emphasis) msgid "Python" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml177(link) msgid "http://www.python.org/ftp/python/2.7.3/python-2.7.3.msi" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml181(para) msgid "Install the MSI accepting the default options." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml184(para) msgid "The installation will put python in C:/python27." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml172(para) msgid "" "Python 2.7.3 must be installed prior to installing the OpenStack Compute " "Driver on the Hyper-V server. Download and then install the MSI for windows " "here:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml188(emphasis) msgid "Setuptools" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml189(para) msgid "" "You will require pip to install the necessary python module dependencies. " "The installer will install under the C:\\python27 directory structure. " "Setuptools for Python 2.7 for Windows can be download from here:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml194(link) msgid "" "http://pypi.python.org/packages/2.7/s/setuptools/setuptools-0.6c11.win32-py2.7.exe" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml197(emphasis) msgid "Python Dependencies" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml198(para) msgid "" "You must download and manually install the following packages on the Compute" " node:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml202(emphasis) msgid "MySQL-python" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml205(link) msgid "http://codegood.com/download/10/" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml209(emphasis) msgid "pywin32" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml210(para) msgid "Download and run the installer from the following location" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml214(link) msgid "" "http://sourceforge.net/projects/pywin32/files/pywin32/Build%20217/pywin32-217.win32-py2.7.exe" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml218(emphasis) msgid "greenlet" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml219(para) msgid "Select the link below:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml221(link) msgid "http://www.lfd.uci.edu/~gohlke/pythonlibs/" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml222(para) msgid "" "You must scroll to the greenlet section for the following file: " "greenlet-0.4.0.win32-py2.7.‌exe" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml224(para) msgid "" "Click on the file, to initiate the download. Once the download is complete, " "run the installer." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml228(para) msgid "" "You must install the following Python packages through or " ". Run the following replacing PACKAGENAME with the following" " packages:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml231(replaceable) msgid "PACKAGE_NAME" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml235(para) msgid "amqplib" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml238(para) msgid "anyjson" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml241(para) msgid "distribute" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml244(para) msgid "eventlet" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml247(para) msgid "httplib2" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml250(para) msgid "iso8601" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml253(para) msgid "jsonschema" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml256(para) msgid "kombu" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml259(para) msgid "netaddr" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml262(para) msgid "paste" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml265(para) msgid "paste-deploy" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml268(para) msgid "prettytable" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml271(para) msgid "python-cinderclient" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml274(para) msgid "python-glanceclient" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml277(para) msgid "python-keystoneclient" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml280(para) msgid "repoze.lru" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml283(para) msgid "routes" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml286(para) msgid "sqlalchemy" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml289(para) msgid "simplejson" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml292(para) msgid "warlock" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml295(para) msgid "webob" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml298(para) msgid "wmi" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml303(title) msgid "Install Nova-compute" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml304(emphasis) msgid "Using git on Windows to retrieve source" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml305(para) msgid "" "Git be used to download the necessary source code. The installer to run Git " "on Windows can be downloaded here:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml310(link) msgid "" "http://code.google.com/p/msysgit/downloads/list?q=full+installer+official+git" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml312(para) msgid "" "Download the latest installer. Once the download is complete double click " "the installer and follow the prompts in the installation wizard. The default" " should be acceptable for the needs of the document." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml315(para) msgid "Once installed you may run the following to clone the Nova code." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml321(title) msgid "Configure Nova.conf" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml322(para) msgid "" "The nova.conf file must be placed in " "C:\\etc\\nova for running OpenStack on Hyper-V. Below is " "a sample nova.conf for Windows:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml346(para) msgid "The following table contains a reference of all options for hyper-v" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml350(title) msgid "Prepare images for use with Hyper-V" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml351(para) msgid "" "Hyper-V currently supports only the VHD file format for virtual machine " "instances. Detailed instructions for installing virtual machines on Hyper-V " "can be found here:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml355(link) msgid "http://technet.microsoft.com/en-us/library/cc772480.aspx" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml356(para) msgid "" "Once you have successfully created a virtual machine, you can then upload " "the image to glance using the native glance-client:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml359(replaceable) msgid "VM_IMAGE_NAME" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml363(title) msgid "Run Compute with Hyper-V" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml364(para) msgid "" "To start the nova-compute " "service, run this command from a console in the Windows server:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml371(title) msgid "Troubleshoot Hyper-V configuration" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml374(para) msgid "" "I ran the nova-manage service list command from my " "controller; however, I'm not seeing smiley faces for Hyper-V compute nodes, " "what do I do?" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml379(link) msgid "here" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_hyper-v.xml377(emphasis) msgid "" "Verify that you are synchronized with a network time source. Instructions " "for configuring NTP on your Hyper-V compute node are located " "" msgstr "" #. When image changes, this message will be marked fuzzy or untranslated for #. you. #. It doesn't matter what you translate it to: it's not used at all. #: ./doc/config-reference/compute/section_hypervisor_vmware.xml39(None) msgid "" "@@image: '../../common/figures/vmware-nova-driver-architecture.jpg'; " "md5=d95084ce963cffbe3e86307c87d804c1" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml6(title) msgid "VMware vSphere" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml9(title) msgid "Introduction" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml10(para) msgid "" "OpenStack Compute supports the VMware vSphere product family and enables " "access to advanced features such as vMotion, High Availability, and Dynamic " "Resource Scheduling (DRS). This section describes how to configure VMware-" "based virtual machine images for launch. vSphere versions 4.1 and newer are " "supported." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml16(para) msgid "" "The VMware vCenter driver enables the nova-" "compute service to communicate with a VMware vCenter server " "that manages one or more ESX host clusters. The driver aggregates the ESX " "hosts in each cluster to present one large hypervisor entity for each " "cluster to the Compute scheduler. Because individual ESX hosts are not " "exposed to the scheduler, Compute schedules to the granularity of clusters " "and vCenter uses DRS to select the actual ESX host within the cluster. When " "a virtual machine makes its way into a vCenter cluster, it can use all " "vSphere features." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml26(para) msgid "" "The following sections describe how to configure the VMware vCenter driver." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml30(title) msgid "High-level architecture" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml31(para) msgid "" "The following diagram shows a high-level view of the VMware driver " "architecture:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml34(title) msgid "VMware driver architecture" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml43(para) msgid "" "As the figure shows, the OpenStack Compute Scheduler sees three hypervisors " "that each correspond to a cluster in vCenter. Nova-compute contains the VMware driver. You can run with " "multiple nova-compute services. " "While Compute schedules at the granularity of a cluster, the VMware driver " "inside nova-compute interacts " "with the vCenter APIs to select an appropriate ESX host within the cluster. " "Internally, vCenter uses DRS for placement." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml53(para) msgid "" "The VMware vCenter driver also interacts with the OpenStack Image Service to" " copy VMDK images from the Image Service back end store. The dotted line in " "the figure represents VMDK images being copied from the OpenStack Image " "Service to the vSphere data store. VMDK images are cached in the data store " "so the copy operation is only required the first time that the VMDK image is" " used." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml60(para) msgid "" "After OpenStack boots a VM into a vSphere cluster, the VM becomes visible in" " vCenter and can access vSphere advanced features. At the same time, the VM " "is visible in the OpenStack dashboard and you can manage it as you would any" " other OpenStack VM. You can perform advanced vSphere operations in vCenter " "while you configure OpenStack resources such as VMs through the OpenStack " "dashboard." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml67(para) msgid "" "The figure does not show how networking fits into the architecture. Both " "nova-network and the OpenStack " "Networking Service are supported. For details, see ." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml74(title) msgid "Configuration overview" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml75(para) msgid "" "To get started with the VMware vCenter driver, complete the following high-" "level steps:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml79(para) msgid "Configure vCenter correctly. See ." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml83(para) msgid "" "Configure nova.conf for the VMware vCenter driver. See " "." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml88(para) msgid "" "Load desired VMDK images into the OpenStack Image Service. See ." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml92(para) msgid "" "Configure networking with either nova-" "network or the OpenStack Networking Service. See ." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml100(title) msgid "Prerequisites and limitations" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml101(para) msgid "" "Use the following list to prepare a vSphere environment that runs with the " "VMware vCenter driver:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml105(para) msgid "" "vCenter inventory. Make sure that any " "vCenter used by OpenStack contains a single data center. A future Havana " "stable release will address this temporary limitation." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml111(para) msgid "" "DRS. For any cluster that contains " "multiple ESX hosts, enable DRS and enable fully automated placement." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml116(para) msgid "" "Shared storage. Only shared storage is " "supported and data stores must be shared among all hosts in a cluster. It is" " recommended to remove data stores not intended for OpenStack from clusters " "being configured for OpenStack. Currently, a single data store can be used " "per cluster. A future Havana stable release will address this temporary " "limitation." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml125(para) msgid "" "Clusters and data stores. Do not use " "OpenStack clusters and data stores for other purposes. If you do, OpenStack " "displays incorrect usage information." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml131(para) msgid "" "Networking. The networking configuration " "depends on the desired networking model. See ." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml136(para) msgid "" "Security groups. If you use the VMware " "driver with OpenStack Networking and the NSX plug-in, security groups are " "supported. If you use nova-" "network, security groups are not supported." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml141(para) msgid "The NSX plug-in is the only plug-in that is validated for vSphere." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml145(para) msgid "" "VNC. The port range 5900 - 6105 " "(inclusive) is automatically enabled for VNC connections on every ESX host " "in all clusters under OpenStack control. For more information about using a " "VNC client to connect to virtual machine, see http://kb.vmware.com/selfservice/microsites/search.do?language=en_US&cmd=displayKC&externalId=1246." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml152(para) msgid "" "In addition to the default VNC port numbers (5900 to 6000) specified in the " "above document, the following ports are also used: 6101, 6102, and 6105." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml155(para) msgid "" "You must modify the ESXi firewall configuration to allow the VNC ports. " "Additionally, for the firewall modifications to persist after a reboot, you " "must create a custom vSphere Installation Bundle (VIB) which is then " "installed onto the running ESXi host or added to a custom image profile used" " to install ESXi hosts. For details about how to create a VIB for persisting" " the firewall configuration modifications, see " " " "http://kb.vmware.com/selfservice/microsites/search.do?language=en_US&cmd=displayKC&externalId=2007381." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml168(para) msgid "" "Ephemeral Disks. Ephemeral disks are not " "supported. A future stable release will address this temporary limitation." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml175(title) msgid "VMware vCenter driver" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml176(para) msgid "" "Use the VMware vCenter driver (VMwareVCDriver) to connect OpenStack Compute " "with vCenter. This recommended configuration enables access through vCenter " "to advanced vSphere features like vMotion, High Availability, and Dynamic " "Resource Scheduling (DRS)." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml182(title) msgid "VMwareVCDriver configuration options" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml183(para) msgid "" "When you use the VMwareVCDriver (vCenter) with OpenStack Compute, add the " "following VMware-specific configuration options to the " "nova.conf file:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml200(para) msgid "" "Clusters: The vCenter driver can support multiple clusters. To use more than" " one cluster, simply add multiple cluster_name lines in " "nova.conf with the appropriate cluster name. Clusters " "and data stores used by the vCenter driver should not contain any VMs other " "than those created by the driver." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml209(para) msgid "" "Data stores: The datastore_regex field specifies the data " "stores to use with Compute. For example, " "datastore_regex=\"nas.*\" selects all the data stores that have" " a name starting with \"nas\". If this line is omitted, Compute uses the " "first data store returned by the vSphere API. It is recommended not to use " "this field and instead remove data stores that are not intended for " "OpenStack." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml220(para) msgid "" "A nova-compute service can " "control one or more clusters containing multiple ESX hosts, making " "nova-compute a critical service " "from a high availability perspective. Because the host that runs nova-compute can fail while the vCenter and " "ESX still run, you must protect the nova-" "compute service against host failures." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml230(para) #: ./doc/config-reference/compute/section_hypervisor_vmware.xml634(para) msgid "" "Many nova.conf options are relevant to libvirt but do " "not apply to this driver." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml233(para) msgid "" "You must complete additional configuration for environments that use vSphere" " 5.0 and earlier. See ." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml239(title) msgid "Images with VMware vSphere" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml240(para) msgid "" "The vCenter driver supports images in the VMDK format. Disks in this format " "can be obtained from VMware Fusion or from an ESX environment. It is also " "possible to convert other formats, such as qcow2, to the VMDK format using " "the qemu-img utility. After a VMDK disk is available, load it " "into the OpenStack Image Service. Then, you can use it with the VMware " "vCenter driver. The following sections provide additional details on the " "supported disks and the commands used for conversion and upload." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml250(title) msgid "Supported image types" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml251(para) msgid "" "Upload images to the OpenStack Image Service in VMDK format. The following " "VMDK disk types are supported:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml255(para) msgid "" "VMFS Flat Disks (includes thin, thick, " "zeroedthick, and eagerzeroedthick). Note that once a VMFS thin disk is " "exported from VMFS to a non-VMFS location, like the OpenStack Image Service," " it becomes a preallocated flat disk. This impacts the transfer time from " "the OpenStack Image Service to the data store when the full preallocated " "flat disk, rather than the thin disk, must be transferred." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml265(para) msgid "" "Monolithic Sparse disks. Sparse disks " "get imported from the OpenStack Image Service into ESX as thin provisioned " "disks. Monolithic Sparse disks can be obtained from VMware Fusion or can be " "created by converting from other virtual disk formats using the qemu-" "img utility." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml274(para) msgid "" "The following table shows the vmware_disktype property that " "applies to each of the supported VMDK disk types:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml278(caption) msgid "OpenStack Image Service disk type settings" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml281(th) msgid "vmware_disktype property" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml282(th) #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml31(title) msgid "VMDK disk type" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml287(td) msgid "sparse" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml289(para) msgid "Monolithic Sparse" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml293(td) #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml51(td) #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml53(td) msgid "thin" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml295(para) msgid "VMFS flat, thin provisioned" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml299(td) msgid "preallocated (default)" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml301(para) msgid "VMFS flat, thick/zeroedthick/eagerzeroedthick" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml308(para) msgid "" "The vmware_disktype property is set when an image is loaded " "into the OpenStack Image Service. For example, the following command creates" " a Monolithic Sparse image by setting vmware_disktype to " "sparse:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml317(para) msgid "" "Note that specifying thin does not provide any advantage " "over preallocated with the current version of the driver." " Future versions might restore the thin properties of the disk after it is " "downloaded to a vSphere data store." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml324(title) msgid "Convert and load images" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml325(para) msgid "" "Using the qemu-img utility, disk images in several formats " "(such as, qcow2) can be converted to the VMDK format." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml328(para) msgid "" "For example, the following command can be used to convert a qcow2 Ubuntu Precise cloud image:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml334(para) msgid "" "VMDK disks converted through qemu-img are always monolithic sparse VMDK disks with an IDE " "adapter type. Using the previous example of the Precise Ubuntu image after " "the qemu-img conversion, the command to upload the VMDK disk " "should be something like:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml345(para) msgid "" "Note that the vmware_disktype is set to sparse and the vmware_adaptertype is" " set to ide in the previous command." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml349(para) msgid "" "If the image did not come from the qemu-img utility, the " "vmware_disktype and vmware_adaptertype might be " "different. To determine the image adapter type from an image file, use the " "following command and look for the ddb.adapterType= line:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml356(para) msgid "" "Assuming a preallocated disk type and an iSCSI lsiLogic adapter type, the " "following command uploads the VMDK disk:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml364(para) msgid "" "Currently, OS boot VMDK disks with an IDE adapter type cannot be attached to" " a virtual SCSI controller and likewise disks with one of the SCSI adapter " "types (such as, busLogic, lsiLogic) cannot be attached to the IDE " "controller. Therefore, as the previous examples show, it is important to set" " the vmware_adaptertype property correctly. The default adapter" " type is lsiLogic, which is SCSI, so you can omit the " "vmware_adaptertype property if you are certain that " "the image adapter type is lsiLogic." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml376(title) msgid "Tag VMware images" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml377(para) msgid "" "In a mixed hypervisor environment, OpenStack Compute uses the " "hypervisor_type tag to match images to the correct hypervisor " "type. For VMware images, set the hypervisor type to " "vmware. Other valid hypervisor types include: xen, qemu, " "kvm, lxc, uml, and hyperv." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml390(title) msgid "Optimize images" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml391(para) msgid "" "Monolithic Sparse disks are considerably faster to download but have the " "overhead of an additional conversion step. When imported into ESX, sparse " "disks get converted to VMFS flat thin provisioned disks. The download and " "conversion steps only affect the first launched instance that uses the " "sparse disk image. The converted disk image is cached, so subsequent " "instances that use this disk image can simply use the cached version." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml399(para) msgid "" "To avoid the conversion step (at the cost of longer download times) consider" " converting sparse disks to thin provisioned or preallocated disks before " "loading them into the OpenStack Image Service. Below are some tools that can" " be used to pre-convert sparse disks." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml406(emphasis) msgid "Using vSphere CLI (or sometimes called the remote CLI or rCLI) tools" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml408(para) msgid "" "Assuming that the sparse disk is made available on a data store accessible " "by an ESX host, the following command converts it to preallocated format:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml412(para) msgid "" "(Note that the vifs tool from the same CLI package can be used to upload the" " disk to be converted. The vifs tool can also be used to download the " "converted disk if necessary.)" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml418(emphasis) msgid "Using vmkfstools directly on the ESX host" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml420(para) msgid "" "If the SSH service is enabled on an ESX host, the sparse disk can be " "uploaded to the ESX data store via scp and the vmkfstools local to the ESX " "host can use used to perform the conversion: (After logging in to the host " "via ssh)" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml429(emphasis) msgid "vmware-vdiskmanager" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml430(para) msgid "" "vmware-vdiskmanager is a utility that comes bundled with VMware" " Fusion and VMware Workstation. Below is an example of converting a sparse " "disk to preallocated format:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml435(para) msgid "" "In all of the above cases, the converted vmdk is actually a pair of files: " "the descriptor file converted.vmdk and " "the actual virtual disk data file converted-" "flat.vmdk. The file to be uploaded to the OpenStack Image Service" " is converted-flat.vmdk." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml446(title) msgid "Image handling" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml447(para) msgid "" "The ESX hypervisor requires a copy of the VMDK file in order to boot up a " "virtual machine. As a result, the vCenter OpenStack Compute driver must " "download the VMDK via HTTP from the OpenStack Image Service to a data store " "that is visible to the hypervisor. To optimize this process, the first time " "a VMDK file is used, it gets cached in the data store. Subsequent virtual " "machines that need the VMDK use the cached version and don't have to copy " "the file again from the OpenStack Image Service." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml456(para) msgid "" "Even with a cached VMDK, there is still a copy operation from the cache " "location to the hypervisor file directory in the shared data store. To avoid" " this copy, boot the image in linked_clone mode. To learn how to enable this" " mode, see . Note also that it is possible " "to override the linked_clone mode on a per-image basis by using the " "vmware_linked_clone property in the OpenStack Image Service." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml467(title) msgid "Networking with VMware vSphere" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml468(para) msgid "" "The VMware driver supports networking with the nova-network service or the OpenStack Networking Service. " "Depending on your installation, complete these configuration steps before " "you provision VMs:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml475(para) msgid "" "The nova-" "network service with the FlatManager or " "FlatDHCPManager. Create a port group with the same name as the " "flat_network_bridge value in the " "nova.conf file. The default value is " "br100." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml481(para) msgid "All VM NICs are attached to this port group." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml482(para) msgid "" "Ensure that the flat interface of the node that runs the nova-network service has a path to this " "network." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml487(para) msgid "" "The nova-" "network service with the VlanManager. Set the " "vlan_interface configuration option to match the ESX host" " interface that handles VLAN-tagged VM traffic." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml493(para) msgid "OpenStack Compute automatically creates the corresponding port groups." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml497(para) msgid "" "The OpenStack Networking Service. If you " "use OVS as the l2 agent, create a port group with the " "same name as the DEFAULT.neutron_ovs_bridge value in the " "nova.conf file. Otherwise, create a port group with the" " same name as the vmware.integration_bridge value in the " "nova.conf file. In both cases, the default value is " "br-int." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml506(para) msgid "" "All VM NICs are attached to this port group for management by the OpenStack " "Networking plug-in." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml512(title) msgid "Volumes with VMware vSphere" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml513(para) msgid "" "The VMware driver supports attaching volumes from the OpenStack Block " "Storage service. The VMware VMDK driver for OpenStack Block Storage is " "recommended and should be used for managing volumes based on vSphere data " "stores. More information about the VMware VMDK driver can be found at: VMware VMDK Driver. Also an iscsi volume driver " "provides limited support and can be used only for attachments." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml524(title) msgid "vSphere 5.0 and earlier additional set up" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml525(para) msgid "" "Users of vSphere 5.0 or earlier must host their WSDL files locally. These " "steps are applicable for vCenter 5.0 or ESXi 5.0 and you can either mirror " "the WSDL from the vCenter or ESXi server that you intend to use or you can " "download the SDK directly from VMware. These workaround steps fix a known" " issue with the WSDL that was resolved in later versions." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml534(title) msgid "Mirror WSDL from vCenter (or ESXi)" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml536(para) msgid "" "Set the VMWAREAPI_IP shell variable to the IP address for your " "vCenter or ESXi host from where you plan to mirror files. For example:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml542(para) msgid "Create a local file system directory to hold the WSDL files:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml547(para) msgid "Change into the new directory. " msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml551(para) msgid "" "Use your OS-specific tools to install a command-line tool that can download " "files like ." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml556(para) msgid "Download the files to the local file cache:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml566(para) msgid "" "Because the reflect-types.xsd and reflect-" "messagetypes.xsd files do not fetch properly, you must stub out " "these files. Use the following XML listing to replace the missing file " "content. The XML parser underneath Python can be very particular and if you " "put a space in the wrong place, it can break the parser. Copy the following " "contents and formatting carefully." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml583(para) msgid "" "Now that the files are locally present, tell the driver to look for the SOAP" " service WSDLs in the local file system and not on the remote vSphere " "server. Add the following setting to the nova.conf file" " for your nova-compute node:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml593(para) msgid "" "Alternatively, download the version appropriate SDK from http://www.vmware.com/support/developer/vc-sdk/ and copy it to" " the /opt/stack/vmware file. Make sure that the WSDL is" " available, in for example " "/opt/stack/vmware/SDK/wsdl/vim25/vimService.wsdl. You " "must point nova.conf to fetch this WSDL file from the " "local file system by using a URL." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml602(para) msgid "" "When using the VMwareVCDriver (vCenter) with OpenStack Compute with vSphere " "version 5.0 or earlier, nova.conf must include the " "following extra config option:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml610(title) msgid "VMware ESX driver" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml611(para) msgid "" "This section covers details of using the VMwareESXDriver. The ESX Driver has" " not been extensively tested and is not recommended. To configure the VMware" " vCenter driver instead, see ." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml616(title) msgid "VMwareESXDriver configuration options" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml617(para) msgid "" "When you use the VMwareESXDriver (no vCenter) with OpenStack Compute, add " "the following VMware-specific configuration options to the " "nova.conf file:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml629(para) msgid "" "Remember that you will have one nova-" "compute service for each ESXi host. It is recommended that this" " host run as a VM on the same ESXi host that it manages." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml639(title) msgid "Requirements and limitations" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml640(para) msgid "" "The ESXDriver cannot use many of the vSphere platform advanced capabilities," " namely vMotion, high availability, and DRS." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_vmware.xml646(title) #: ./doc/config-reference/compute/section_compute-scheduler.xml527(title) msgid "Configuration reference" msgstr "" #. When image changes, this message will be marked fuzzy or untranslated for #. you. #. It doesn't matter what you translate it to: it's not used at all. #: ./doc/config-reference/compute/section_compute-scheduler.xml75(None) msgid "" "@@image: '../../common/figures/filteringWorkflow1.png'; " "md5=c144af5cbdee1bd17a7bde0bea5b5fe7" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml10(title) msgid "Scheduling" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml11(para) msgid "" "Compute uses the nova-scheduler " "service to determine how to dispatch compute and volume requests. For " "example, the nova-scheduler " "service determines which host a VM should launch on. The term " "host in the context of filters means a physical node " "that has a nova-compute service " "running on it. You can configure the scheduler through a variety of options." msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml17(para) msgid "Compute is configured with the following default scheduler options:" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml24(para) msgid "" "By default, the compute scheduler is configured as a filter scheduler, as " "described in the next section. In the default configuration, this scheduler " "considers hosts that meet all the following criteria:" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml30(para) msgid "" "Are in the requested availability zone " "(AvailabilityZoneFilter)." msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml34(para) msgid "Have sufficient RAM available (RamFilter)." msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml38(para) msgid "" "Are capable of servicing the request (ComputeFilter)." msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml42(para) msgid "" "For information on the volume scheduler, refer the Block Storage section of " "OpenStack Cloud Administrator " "Guide for information." msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml46(title) msgid "Filter scheduler" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml47(para) msgid "" "The Filter Scheduler " "(nova.scheduler.filter_scheduler.FilterScheduler) is the " "default scheduler for scheduling virtual machine instances. It supports " "filtering and weighting to make informed decisions on where a new instance " "should be created. You can use this scheduler to schedule compute requests " "but not volume requests. For example, you can use it with only the " "compute_scheduler_driver configuration option." msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml60(title) msgid "Filters" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml70(title) msgid "Filtering" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml61(para) msgid "" "When the Filter Scheduler receives a request for a resource, it first " "applies filters to determine which hosts are eligible for consideration when" " dispatching a resource. Filters are binary: either a host is accepted by " "the filter, or it is rejected. Hosts that are accepted by the filter are " "then processed by a different algorithm to decide which hosts to use for " "that request, described in the Weights " "section. " msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml80(para) msgid "" "The scheduler_available_filters configuration option in " "nova.conf provides the Compute service with the list of" " the filters that are used by the scheduler. The default setting specifies " "all of the filter that are included with the Compute service:" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml87(para) msgid "" "This configuration option can be specified multiple times. For example, if " "you implemented your own custom filter in Python called " "myfilter.MyFilter and you wanted to use both the built-in" " filters and your custom filter, your nova.conf file " "would contain:" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml96(para) msgid "" "The scheduler_default_filters configuration option in " "nova.conf defines the list of filters that are applied " "by the nova-scheduler service. As" " mentioned, the default filters are:" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml102(para) msgid "The following sections describe the available filters." msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml105(title) msgid "AggregateCoreFilter" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml106(para) msgid "" "Implements blueprint per-aggregate-resource-ratio. AggregateCoreFilter " "supports per-aggregate cpu_allocation_ratio. If the per-" "aggregate value is not found, the value falls back to the global setting." msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml113(title) msgid "AggregateInstanceExtraSpecsFilter" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml114(para) msgid "" "Matches properties defined in an instance type's extra specs against admin-" "defined properties on a host aggregate. Works with specifications that are " "unscoped, or are scoped with " "aggregate_instance_extra_specs. See the host aggregates section for documentation on how" " to use this filter." msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml123(title) msgid "AggregateMultiTenancyIsolation" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml124(para) msgid "" "Isolates tenants to specific host " "aggregates. If a host is in an aggregate that has the metadata key " "filter_tenant_id it only creates instances from that " "tenant (or list of tenants). A host can be in different aggregates. If a " "host does not belong to an aggregate with the metadata key, it can create " "instances from all tenants." msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml134(title) msgid "AggregateRamFilter" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml135(para) msgid "" "Implements blueprint per-aggregate-resource-ratio. " "Supports per-aggregate ram_allocation_ratio. If per-" "aggregate value is not found, it falls back to the default setting." msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml143(title) msgid "AllHostsFilter" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml144(para) msgid "" "This is a no-op filter, it does not eliminate any of the available hosts." msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml148(title) msgid "AvailabilityZoneFilter" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml149(para) msgid "" "Filters hosts by availability zone. This filter must be enabled for the " "scheduler to respect availability zones in requests." msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml154(title) msgid "ComputeCapabilitiesFilter" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml155(para) msgid "" "Matches properties defined in an instance type's extra specs against compute" " capabilities." msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml157(para) msgid "" "If an extra specs key contains a colon \":\", anything before the colon is " "treated as a namespace, and anything after the colon is treated as the key " "to be matched. If a namespace is present and is not 'capabilities', it is " "ignored by this filter." msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml163(para) msgid "" "Disable the ComputeCapabilitiesFilter when using a Bare Metal configuration," " due to bug " "1129485" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml170(title) msgid "ComputeFilter" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml171(para) msgid "Passes all hosts that are operational and enabled." msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml173(para) msgid "In general, this filter should always be enabled." msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml177(title) msgid "CoreFilter" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml178(para) msgid "" "Only schedule instances on hosts if there are sufficient CPU cores " "available. If this filter is not set, the scheduler may over provision a " "host based on cores (for example, the virtual cores running on an instance " "may exceed the physical cores)." msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml183(para) msgid "" "This filter can be configured to allow a fixed amount of vCPU overcommitment" " by using the cpu_allocation_ratio Configuration option " "in nova.conf. The default setting is:" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml190(para) msgid "" "With this setting, if 8 vCPUs are on a node, the scheduler allows instances " "up to 128 vCPU to be run on that node." msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml193(para) msgid "To disallow vCPU overcommitment set:" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml197(title) msgid "DifferentHostFilter" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml198(para) msgid "" "Schedule the instance on a different host from a set of instances. To take " "advantage of this filter, the requester must pass a scheduler hint, using " "different_host as the key and a list of instance uuids as" " the value. This filter is the opposite of the " "SameHostFilter. Using the command-line " "tool, use the --hint flag. For example:" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml210(para) msgid "" "With the API, use the os:scheduler_hints key. For " "example:" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml227(title) msgid "DiskFilter" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml228(para) msgid "" "Only schedule instances on hosts if there is sufficient disk space available" " for root and ephemeral storage." msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml231(para) msgid "" "This filter can be configured to allow a fixed amount of disk overcommitment" " by using the disk_allocation_ratio Configuration option " "in nova.conf. The default setting is:" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml238(para) msgid "" "Adjusting this value to greater than 1.0 enables scheduling instances while " "over committing disk resources on the node. This might be desirable if you " "use an image format that is sparse or copy on write such that each virtual " "instance does not require a 1:1 allocation of virtual disk to physical " "storage." msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml245(title) msgid "GroupAffinityFilter" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml246(para) msgid "" "The GroupAffinityFilter ensures that an instance is scheduled on to a host " "from a set of group hosts. To take advantage of this filter, the requester " "must pass a scheduler hint, using group as the key and an" " arbitrary name as the value. Using the command-line tool, " "use the --hint flag. For example:" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml258(title) msgid "GroupAntiAffinityFilter" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml259(para) msgid "" "The GroupAntiAffinityFilter ensures that each instance in a group is on a " "different host. To take advantage of this filter, the requester must pass a " "scheduler hint, using group as the key and an arbitrary " "name as the value. Using the command-line tool, use the " "--hint flag. For example:" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml271(title) msgid "ImagePropertiesFilter" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml272(para) msgid "" "Filters hosts based on properties defined on the instance's image. It passes" " hosts that can support the specified image properties contained in the " "instance. Properties include the architecture, hypervisor type, and virtual " "machine mode. for example, an instance might require a host that runs an " "ARM-based processor and QEMU as the hypervisor. An image can be decorated " "with these properties by using:" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml281(para) msgid "The image properties that the filter checks for are:" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml285(para) msgid "" "architecture: Architecture describes the machine " "architecture required by the image. Examples are i686, x86_64, arm, and " "ppc64." msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml291(para) msgid "" "hypervisor_type: Hypervisor type describes the hypervisor" " required by the image. Examples are xen, kvm, qemu, and xenapi." msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml297(para) msgid "" "vm_mode: Virtual machine mode describes the hypervisor " "application binary interface (ABI) required by the image. Examples are 'xen'" " for Xen 3.0 paravirtual ABI, 'hvm' for native ABI, 'uml' for User Mode " "Linux paravirtual ABI, exe for container virt executable ABI." msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml308(title) msgid "IsolatedHostsFilter" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml309(para) msgid "" "Allows the admin to define a special (isolated) set of images and a special " "(isolated) set of hosts, such that the isolated images can only run on the " "isolated hosts, and the isolated hosts can only run isolated images. The " "flag restrict_isolated_hosts_to_isolated_images can be " "used to force isolated hosts to only run isolated images." msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml315(para) msgid "" "The admin must specify the isolated set of images and hosts in the " "nova.conf file using the " "isolated_hosts and isolated_images " "configuration options. For example: " msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml325(title) msgid "JsonFilter" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml330(para) msgid "=" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml333(para) msgid "<" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml336(para) msgid ">" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml339(para) msgid "in" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml342(para) msgid "<=" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml345(para) msgid ">=" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml348(para) msgid "not" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml351(para) msgid "or" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml354(para) msgid "and" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml358(para) msgid "$free_ram_mb" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml361(para) msgid "$free_disk_mb" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml364(para) msgid "$total_usable_ram_mb" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml367(para) msgid "$vcpus_total" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml370(para) msgid "$vcpus_used" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml326(para) msgid "" "The JsonFilter allows a user to construct a custom filter by passing a " "scheduler hint in JSON format. The following operators are " "supported:The filter supports the following " "variables:Using the command-line tool, use " "the --hint flag:" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml377(para) #: ./doc/config-reference/compute/section_compute-scheduler.xml436(para) #: ./doc/config-reference/compute/section_compute-scheduler.xml482(para) msgid "With the API, use the os:scheduler_hints key:" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml393(title) msgid "RamFilter" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml394(para) msgid "" "Only schedule instances on hosts that have sufficient RAM available. If this" " filter is not set, the scheduler may over provision a host based on RAM " "(for example, the RAM allocated by virtual machine instances may exceed the " "physical RAM)." msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml399(para) msgid "" "This filter can be configured to allow a fixed amount of RAM overcommitment " "by using the ram_allocation_ratio configuration option in" " nova.conf. The default setting is:" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml406(para) msgid "" "With this setting, if there is 1GB of free RAM, the scheduler allows " "instances up to size 1.5GB to be run on that instance." msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml411(title) msgid "RetryFilter" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml412(para) msgid "" "Filter out hosts that have already been attempted for scheduling purposes. " "If the scheduler selects a host to respond to a service request, and the " "host fails to respond to the request, this filter prevents the scheduler " "from retrying that host for the service request." msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml417(para) msgid "" "This filter is only useful if the scheduler_max_attempts " "configuration option is set to a value greater than zero." msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml423(title) msgid "SameHostFilter" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml424(para) msgid "" "Schedule the instance on the same host as another instance in a set of " "instances. To take advantage of this filter, the requester must pass a " "scheduler hint, using same_host as the key and a list of " "instance uuids as the value. This filter is the opposite of the " "DifferentHostFilter. Using the command-" "line tool, use the --hint flag:" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml453(title) msgid "SimpleCIDRAffinityFilter" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml454(para) msgid "" "Schedule the instance based on host IP subnet range. To take advantage of " "this filter, the requester must specify a range of valid IP address in CIDR " "format, by passing two scheduler hints:" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml460(literal) msgid "build_near_host_ip" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml462(para) msgid "" "The first IP address in the subnet (for example, " "192.168.1.1)" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml468(literal) msgid "cidr" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml470(para) msgid "" "The CIDR that corresponds to the subnet (for example, " "/24)" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml476(para) msgid "" "Using the command-line tool, use the " "--hint flag. For example, to specify the IP subnet " "192.168.1.1/24" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml499(title) msgid "Weights" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml501(para) msgid "" "The Filter Scheduler weighs hosts based on the config option " "scheduler_weight_classes, this defaults to " "nova.scheduler.weights.all_weighers, which selects the " "only weigher available -- the RamWeigher. Hosts are then weighed and sorted " "with the largest weight winning." msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml510(para) msgid "" "The default is to spread instances across all hosts evenly. Set the " "ram_weight_multiplier option to a negative number if you " "prefer stacking instead of spreading." msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml516(title) msgid "Chance scheduler" msgstr "" #: ./doc/config-reference/compute/section_compute-scheduler.xml518(para) msgid "" "As an administrator, you work with the Filter Scheduler. However, the " "Compute service also uses the Chance Scheduler, " "nova.scheduler.chance.ChanceScheduler, which randomly " "selects from lists of filtered hosts. It is the default volume scheduler." msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml10(title) msgid "Cells" msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml12(para) msgid "" "Cells functionality allows you to scale" " an OpenStack Compute cloud in a more distributed fashion without having to " "use complicated technologies like database and message queue clustering. It " "is intended to support very large deployments." msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml16(para) msgid "" "When this functionality is enabled, the hosts in an OpenStack Compute cloud " "are partitioned into groups called cells. Cells are configured as a tree. " "The top-level cell should have a host that runs a nova-api service, but no nova-compute services. Each child cell should" " run all of the typical nova-* " "services in a regular Compute cloud except for nova-api. You can think of cells as a normal Compute " "deployment in that each cell has its own database server and message queue " "broker." msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml24(para) msgid "" "The nova-cells service handles " "communication between cells and selects cells for new instances. This " "service is required for every cell. Communication between cells is " "pluggable, and currently the only option is communication through RPC." msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml28(para) msgid "" "Cells scheduling is separate from host scheduling. nova-cells first picks a cell (now randomly, " "but future releases plan to add filtering/weighing functionality, and " "decisions will be based on broadcasts of capacity/capabilities). Once a cell" " is selected and the new build request reaches its nova-cells service, it is sent over to the " "host scheduler in that cell and the build proceeds as it would have without " "cells." msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml35(para) msgid "Cell functionality is currently considered experimental." msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml38(title) msgid "Cell configuration options" msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml43(literal) msgid "enable" msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml45(para) msgid "" "Set this is True to turn on cell functionality, which is " "off by default." msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml50(literal) msgid "name" msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml52(para) msgid "Name of the current cell. This must be unique for each cell." msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml56(literal) msgid "capabilities" msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml58(para) msgid "" "List of arbitrary " "key=value" " pairs defining capabilities of the current cell. Values include " "hypervisor=xenserver;kvm,os=linux;windows." msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml66(literal) msgid "call_timeout" msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml68(para) msgid "How long in seconds to wait for replies from calls between cells." msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml73(term) msgid "scheduler_filter_classes" msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml75(para) msgid "" "Filter classes that the cells scheduler should use. By default, uses " "\"nova.cells.filters.all_filters\" to map to all cells " "filters included with Compute." msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml81(term) msgid "scheduler_weight_classes" msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml82(para) msgid "" "Weight classes the cells scheduler should use. By default, uses " "\"nova.cells.weights.all_weighers\" to map to all cells " "weight algorithms (weighers) included with Compute." msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml88(term) msgid "ram_weight_multiplier" msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml90(para) msgid "" "Multiplier used for weighing ram. Negative numbers mean you want Compute to " "stack VMs on one host instead of spreading out new VMs to more hosts in the " "cell. Default value is 10.0." msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml39(para) msgid "" "Cells are disabled by default. All cell-related configuration options go " "under a [cells] section in " "nova.conf. The following cell-related options are " "currently supported:" msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml99(title) msgid "Configure the API (top-level) cell" msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml100(para) msgid "" "The compute API class must be changed in the API cell so that requests can " "be proxied through nova-cells down to the correct cell properly. Add the " "following to nova.conf in the API cell:" msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml112(title) msgid "Configure the child cells" msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml121(replaceable) msgid "cell1" msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml113(para) msgid "" "Add the following to nova.conf in the child cells, " "replacing cell1 with the name of each " "cell:" msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml124(title) msgid "Configure the database in each cell" msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml125(para) msgid "" "Before bringing the services online, the database in each cell needs to be " "configured with information about related cells. In particular, the API cell" " needs to know about its immediate children, and the child cells must know " "about their immediate agents. The information needed is the " "RabbitMQ server credentials for the particular " "cell." msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml130(para) msgid "" "Use the command to add this information to the database in " "each cell:" msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml152(para) msgid "" "As an example, assume we have an API cell named api and a" " child cell named cell1. Within the api cell, we have the" " following RabbitMQ server info:" msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml160(para) msgid "" "And in the child cell named cell1 we have the following " "RabbitMQ server info:" msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml167(para) msgid "We would run this in the API cell, as root." msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml169(para) msgid "Repeat the above for all child cells." msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml170(para) msgid "" "In the child cell, we would run the following, as root:" msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml176(title) msgid "Cell scheduling configuration" msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml177(para) msgid "" "To determine the best cell for launching a new instance, Compute uses a set " "of filters and weights configured in " "/etc/nova/nova.conf. The following options are " "available to prioritize cells for scheduling:" msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml183(para) msgid "" "scheduler_filter_classes - Specifies the list of filter " "classes. By default nova.cells.weights.all_filters is " "specified, which maps to all cells filters included with Compute (see ." msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml189(para) msgid "" "scheduler_weight_classes - Specifies the list of weight " "classes. By default nova.cells.weights.all_weighers is " "specified, which maps to all cell weight algorithms (weighers) included with" " Compute. The following modules are available:" msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml194(para) msgid "" "mute_child: Downgrades the likelihood of child cells being " "chosen for scheduling requests, which haven't sent capacity or capability " "updates in a while. Options include mute_weight_multiplier " "(multiplier for mute children; value should be negative) and " "mute_weight_value (assigned to mute children; should be a " "positive value)." msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml203(para) msgid "" "ram_by_instance_type: Select cells with the most RAM capacity " "for the instance type being requested. Because higher weights win, Compute " "returns the number of available units for the instance type requested. The " "ram_weight_multiplier option defaults to 10.0 that adds to the " "weight by a factor of 10. Use a negative number to stack VMs on one host " "instead of spreading out new VMs to more hosts in the cell." msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml211(para) msgid "" "weight_offset: Allows modifying the database to weight a " "particular cell. You can use this when you want to disable a cell (for " "example, '0'), or to set a default cell by making its weight_offset very " "high (for example, '999999999999999'). The highest weight will be the first " "cell to be scheduled for launching an instance." msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml219(para) msgid "" "Additionally, the following options are available for the cell scheduler:" msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml223(para) msgid "" "scheduler_retries - Specifies how many times the scheduler " "tries to launch a new instance when no cells are available (default=10)." msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml227(para) msgid "" "scheduler_retry_delay - Specifies the delay (in seconds) " "between retries (default=2)." msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml232(para) msgid "" "As an admin user, you can also add a filter that directs builds to a " "particular cell. The policy.json file must have a line " "with \"cells_scheduler_filter:TargetCellFilter\" : " "\"is_admin:True\" to let an admin user specify a scheduler hint to" " direct a build to a particular cell." msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml239(title) msgid "Optional cell configuration" msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml240(para) msgid "" "Cells currently keeps all inter-cell communication data, including user " "names and passwords, in the database. This is undesirable and unnecessary " "since cells data isn't updated very frequently. Instead, create a JSON file " "to input cells data specified via a [cells]cells_config option." " When specified, the database is no longer consulted when reloading cells " "data. The file will need the columns present in the Cell model (excluding " "common database fields and the id column). The queue connection" " information must be specified through a transport_url field, " "instead of username, password, and so on. The " "transport_url has the following form:" msgstr "" #: ./doc/config-reference/compute/section_compute-cells.xml255(para) msgid "" "The scheme can be either qpid or " "rabbit, as shown previously. The following sample shows " "this optional configuration:" msgstr "" #: ./doc/config-reference/compute/section_compute-security.xml10(title) msgid "Security hardening" msgstr "" #: ./doc/config-reference/compute/section_compute-security.xml11(para) msgid "" "OpenStack Compute can be integrated with various third-party technologies to" " increase security. For more information, see the OpenStack Security " "Guide." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_qemu.xml7(title) msgid "QEMU" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_qemu.xml8(para) msgid "" "From the perspective of the Compute service, the QEMU hypervisor is very " "similar to the KVM hypervisor. Both are controlled through libvirt, both " "support the same feature set, and all virtual machine images that are " "compatible with KVM are also compatible with QEMU. The main difference is " "that QEMU does not support native virtualization. Consequently, QEMU has " "worse performance than KVM and is a poor choice for a production deployment." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_qemu.xml15(para) msgid "Running on older hardware that lacks virtualization support." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_qemu.xml19(para) msgid "" "Running the Compute service inside of a virtual machine for development or " "testing purposes, where the hypervisor does not support native " "virtualization for guests." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_qemu.xml13(para) msgid "The typical uses cases for QEMU are" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_qemu.xml25(para) msgid "" "To enable QEMU, add these settings to " "nova.conf:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_qemu.xml29(para) msgid "" "For some operations you may also have to install the " "utility:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_qemu.xml31(para) msgid "On Ubuntu: " msgstr "" #: ./doc/config-reference/compute/section_hypervisor_qemu.xml34(para) msgid "On RHEL, Fedora or CentOS: " msgstr "" #: ./doc/config-reference/compute/section_hypervisor_qemu.xml37(para) msgid "On openSUSE: " msgstr "" #: ./doc/config-reference/compute/section_hypervisor_qemu.xml40(para) msgid "" "The QEMU hypervisor supports the following virtual machine image formats:" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_qemu.xml53(title) msgid "Tips and fixes for QEMU on RHEL" msgstr "" #: ./doc/config-reference/compute/section_hypervisor_qemu.xml54(para) msgid "" "If you are testing OpenStack in a virtual machine, you need to configure " "nova to use qemu without KVM and hardware virtualization. The second command" " relaxes SELinux rules to allow this mode of operation ( " "https://bugzilla.redhat.com/show_bug.cgi?id=753589). The last two " "commands here work around a libvirt issue fixed in RHEL 6.4. Note nested " "virtualization will be the much slower TCG variety, and you should provide " "lots of memory to the top level guest, as the OpenStack-created guests " "default to 2GM RAM with no overcommit." msgstr "" #: ./doc/config-reference/compute/section_hypervisor_qemu.xml65(para) msgid "The second command, , may take a while." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml10(title) msgid "Configure migrations" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml12(para) msgid "" "Only cloud administrators can perform live migrations. If your cloud is " "configured to use cells, you can perform live migration within but not " "between cells." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml16(para) msgid "" "Migration enables an administrator to move a virtual machine instance from " "one compute host to another. This feature is useful when a compute host " "requires maintenance. Migration can also be useful to redistribute the load " "when many VM instances are running on a specific physical machine." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml21(para) msgid "The migration types are:" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml24(para) msgid "" "Migration (or non-live migration). The " "instance is shut down (and the instance knows that it was rebooted) for a " "period of time to be moved to another hypervisor." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml30(para) msgid "" "Live migration (or true live migration). " "Almost no instance downtime. Useful when the instances must be kept running " "during the migration." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml35(para) msgid "The types of live migration are:" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml38(para) msgid "" "Shared storage-based live migration. Both" " hypervisors have access to shared storage." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml43(para) msgid "" "Block live migration. No shared storage " "is required." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml47(para) msgid "" "Volume-backed live migration. When " "instances are backed by volumes rather than ephemeral disk, no shared " "storage is required, and migration is supported (currently only in libvirt-" "based hypervisors)." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml54(para) msgid "" "The following sections describe how to configure your hosts and compute " "nodes for migrations by using the KVM and XenServer hypervisors." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml58(title) msgid "KVM-Libvirt" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml60(title) #: ./doc/config-reference/compute/section_compute-configure-migrations.xml293(title) #: ./doc/config-reference/compute/section_compute-configure-migrations.xml368(title) #: ./doc/config-reference/block-storage/drivers/xen-sm-driver.xml117(title) msgid "Prerequisites" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml62(para) msgid "Hypervisor: KVM with libvirt" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml66(para) msgid "" "Shared storage:NOVA-INST-DIR/instances/ (for example, " "/var/lib/nova/instances) has to be mounted by shared " "storage. This guide uses NFS but other options, including the OpenStack" " Gluster Connector are available." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml75(para) msgid "" "Instances: Instance can be migrated with " "iSCSI based volumes" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml80(title) #: ./doc/config-reference/compute/section_compute-configure-migrations.xml377(title) msgid "Notes" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml83(para) msgid "" "Because the Compute service does not use the libvirt live migration " "functionality by default, guests are suspended before migration and might " "experience several minutes of downtime. For details, see ." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml90(para) msgid "" "This guide assumes the default value for in " "your nova.conf file (NOVA-INST-" "DIR/instances). If you have changed the " "state_path or instances_path " "variables, modify accordingly." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml99(para) msgid "" "You must specify vncserver_listen=0.0.0.0 or live " "migration does not work correctly." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml106(title) msgid "Example Compute installation environment" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml108(para) msgid "" "Prepare at least three servers; for example, HostA, " "HostB, and HostC." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml113(para) msgid "" "HostA is the Cloud Controller, and should run these services: " "nova-api, nova-scheduler, nova-" "network, cinder-volume," " and nova-objectstore." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml123(para) msgid "" "HostB and HostC are the compute nodes that run nova-compute." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml129(para) msgid "" "Ensure that NOVA-INST-DIR (set" " with state_path in the nova.conf " "file) is the same on all hosts." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml136(para) msgid "" "In this example, HostA is the NFSv4 server that exports " "NOVA-INST-DIR/instances, and" " HostB and HostC mount it." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml144(title) msgid "To configure your system" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml146(para) msgid "" "Configure your DNS or /etc/hosts and ensure it is " "consistent across all hosts. Make sure that the three hosts can perform name" " resolution with each other. As a test, use the command to " "ping each host from one another." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml156(para) msgid "" "Ensure that the UID and GID of your nova and libvirt users are identical " "between each of your servers. This ensures that the permissions on the NFS " "mount works correctly." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml162(para) msgid "" "Follow the instructions at the Ubuntu NFS " "HowTo to setup an NFS server on HostA, and NFS Clients on" " HostB and HostC." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml168(para) msgid "" "The aim is to export NOVA-INST-" "DIR/instances from HostA, and " "have it readable and writable by the nova user on HostB " "and HostC." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml175(para) msgid "" "Using your knowledge from the Ubuntu documentation, configure the NFS server" " at HostA by adding this line to the " "/etc/exports file:" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml179(replaceable) #: ./doc/config-reference/compute/section_compute-configure-migrations.xml194(replaceable) #: ./doc/config-reference/compute/section_compute-configure-migrations.xml199(replaceable) #: ./doc/config-reference/compute/section_compute-configure-migrations.xml206(replaceable) #: ./doc/config-reference/compute/section_compute-configure-migrations.xml211(replaceable) msgid "NOVA-INST-DIR" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml180(para) msgid "" "Change the subnet mask (255.255.0.0) to the appropriate " "value to include the IP addresses of HostB and " "HostC. Then restart the NFS server:" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml188(para) msgid "Set the 'execute/search' bit on your shared directory." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml190(para) msgid "" "On both compute nodes, make sure to enable the 'execute/search' bit to allow" " qemu to be able to use the images within the directories. On all hosts, run" " the following command:" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml197(para) msgid "" "Configure NFS at HostB and HostC by adding this line to the " "/etc/fstab file:" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml200(para) msgid "Make sure that you can mount the exported directory can be mounted:" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml203(para) msgid "" "Check that HostA can see the \"NOVA-INST-" "DIR/instances/\" directory:" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml206(filename) #: ./doc/config-reference/compute/section_compute-configure-migrations.xml211(filename) msgid "/instances/" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml208(para) msgid "" "Perform the same check at HostB and HostC, paying special attention to the " "permissions (nova should be able to write):" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml224(para) msgid "" "Update the libvirt configurations. Modify the " "/etc/libvirt/libvirtd.conf file:" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml234(para) msgid "Modify the /etc/libvirt/qemu.conf file:" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml238(para) msgid "Modify the /etc/init/libvirt-bin.conf file:" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml243(para) msgid "Modify the /etc/default/libvirt-bin file:" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml247(para) msgid "" "Restart libvirt. After you run the command, ensure that libvirt is " "successfully restarted:" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml254(para) msgid "Configure your firewall to allow libvirt to communicate between nodes." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml256(para) msgid "" "For information about ports that are used with libvirt, see the " "libvirt documentation By default, libvirt listens on TCP port 16509 " "and an ephemeral TCP range from 49152 to 49261 is used for the KVM " "communications. As this guide has disabled libvirt auth, you should take " "good care that these ports are only open to hosts within your installation." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml267(para) msgid "" "You can now configure options for live migration. In most cases, you do not " "need to configure any options. The following chart is for advanced usage " "only." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml274(title) msgid "Enable true live migration" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml275(para) msgid "" "By default, the Compute service does not use the libvirt live migration " "functionality. To enable this functionality, add the following line to the " "nova.conf file:" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml280(para) msgid "" "The Compute service does not use libvirt's live migration by default because" " there is a risk that the migration process never ends. This can happen if " "the guest operating system dirties blocks on the disk faster than they can " "migrated." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml291(title) msgid "Shared storage" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml295(para) msgid "" "Compatible XenServer hypervisors. For " "more information, see the Requirements" " for Creating Resource Pools section of the XenServer " "Administrator's Guide." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml304(para) msgid "" "Shared storage. An NFS export, visible to" " all XenServer hosts." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml307(para) msgid "" "For the supported NFS versions, see the NFS" " VHD section of the XenServer Administrator's " "Guide." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml314(para) msgid "" "To use shared storage live migration with XenServer hypervisors, the hosts " "must be joined to a XenServer pool. To create that pool, a host aggregate " "must be created with special metadata. This metadata is used by the XAPI " "plug-ins to establish the pool." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml320(title) msgid "To use shared storage live migration with XenServer hypervisors" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml323(para) msgid "" "Add an NFS VHD storage to your master XenServer, and set it as default SR. " "For more information, please refer to the NFS" " VHD section in the XenServer Administrator's " "Guide." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml331(para) msgid "" "Configure all the compute nodes to use the default sr for pool operations. " "Add this line to your nova.conf configuration files " "across your compute nodes:" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml338(para) msgid "Create a host aggregate:" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml340(para) msgid "" "The command displays a table that contains the ID of the newly created " "aggregate." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml342(para) msgid "" "Now add special metadata to the aggregate, to mark it as a hypervisor pool:" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml346(para) msgid "Make the first compute node part of that aggregate:" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml349(para) msgid "At this point, the host is part of a XenServer pool." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml353(para) msgid "Add additional hosts to the pool:" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml356(para) msgid "" "At this point, the added compute node and the host are shut down, to join " "the host to the XenServer pool. The operation fails, if any server other " "than the compute node is running/suspended on your host." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml366(title) msgid "Block migration" msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml370(para) msgid "" "Compatible XenServer hypervisors. The " "hypervisors must support the Storage XenMotion feature. See your XenServer " "manual to make sure your edition has this feature." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml380(para) msgid "" "To use block migration, you must use the --block-" "migrate parameter with the live migration command." msgstr "" #: ./doc/config-reference/compute/section_compute-configure-migrations.xml385(para) msgid "" "Block migration works only with EXT local storage SRs, and the server must " "not have any volumes attached." msgstr "" #: ./doc/config-reference/networking/section_networking-plugins.xml10(title) msgid "Networking plug-ins" msgstr "" #: ./doc/config-reference/networking/section_networking-plugins.xml11(para) msgid "" "OpenStack Networking introduces the concept of a plug-in, which is a back-" "end implementation of the OpenStack Networking API. A plug-in can use a " "variety of technologies to implement the logical API requests. Some " "OpenStack Networking plug-ins might use basic Linux VLANs and IP tables, " "while others might use more advanced technologies, such as L2-in-L3 " "tunneling or OpenFlow. These sections detail the configuration options for " "the various plug-ins." msgstr "" #: ./doc/config-reference/networking/section_networking-plugins.xml22(title) msgid "BigSwitch configuration options" msgstr "" #: ./doc/config-reference/networking/section_networking-plugins.xml28(title) msgid "Brocade configuration options" msgstr "" #: ./doc/config-reference/networking/section_networking-plugins.xml34(title) msgid "CISCO configuration options" msgstr "" #: ./doc/config-reference/networking/section_networking-plugins.xml40(title) msgid "CloudBase Hyper-V plug-in configuration options (deprecated)" msgstr "" #: ./doc/config-reference/networking/section_networking-plugins.xml47(title) msgid "CloudBase Hyper-V Agent configuration options" msgstr "" #: ./doc/config-reference/networking/section_networking-plugins.xml54(title) msgid "Linux bridge plug-in configuration options (deprecated)" msgstr "" #: ./doc/config-reference/networking/section_networking-plugins.xml61(title) msgid "Linux bridge Agent configuration options" msgstr "" #: ./doc/config-reference/networking/section_networking-plugins.xml68(title) msgid "Mellanox configuration options" msgstr "" #: ./doc/config-reference/networking/section_networking-plugins.xml74(title) msgid "Meta Plug-in configuration options" msgstr "" #: ./doc/config-reference/networking/section_networking-plugins.xml75(para) msgid "The Meta Plug-in allows you to use multiple plug-ins at the same time." msgstr "" #: ./doc/config-reference/networking/section_networking-plugins.xml83(title) msgid "MidoNet configuration options" msgstr "" #: ./doc/config-reference/networking/section_networking-plugins.xml89(title) msgid "NEC configuration options" msgstr "" #: ./doc/config-reference/networking/section_networking-plugins.xml95(title) msgid "Nicira NVP configuration options" msgstr "" #: ./doc/config-reference/networking/section_networking-plugins.xml101(title) msgid "Open vSwitch plug-in configuration options (deprecated)" msgstr "" #: ./doc/config-reference/networking/section_networking-plugins.xml108(title) msgid "Open vSwitch Agent configuration options" msgstr "" #: ./doc/config-reference/networking/section_networking-plugins.xml115(title) msgid "PLUMgrid configuration options" msgstr "" #: ./doc/config-reference/networking/section_networking-plugins.xml121(title) msgid "Ryu configuration options" msgstr "" #: ./doc/config-reference/networking/section_networking-plugins-ml2.xml10(title) msgid "Modular Layer 2 (ml2) configuration options" msgstr "" #: ./doc/config-reference/networking/section_networking-plugins-ml2.xml11(para) msgid "" "The Modular Layer 2 (ml2) plug-in has two components: network types and " "mechanisms. You can configure these components separately. This section " "describes these configuration options." msgstr "" #: ./doc/config-reference/networking/section_networking-plugins-ml2.xml16(title) msgid "MTU bug with VXLAN tunnelling" msgstr "" #: ./doc/config-reference/networking/section_networking-plugins-ml2.xml17(para) msgid "" "Due to a bug in Linux Bridge software maximum transmission unit (MTU) " "handling, using VXLAN tunnels does not work by default." msgstr "" #: ./doc/config-reference/networking/section_networking-plugins-ml2.xml22(para) msgid "" "A simple workaround is to increase the MTU value of the physical interface " "and physical switch fabric by at least 50 bytes. For example, increase the " "MTU value to 1550. This value enables an automatic 50-byte MTU difference " "between the physical interface (1500) and the VXLAN interface (automatically" " 1500-50 = 1450). An MTU value of 1450 causes issues when virtual machine " "taps are configured at an MTU value of 1500." msgstr "" #: ./doc/config-reference/networking/section_networking-plugins-ml2.xml33(para) msgid "" "Another workaround is to decrease the virtual ethernet devices' MTU. Set the" " option to 1450 in the " "neutron.conf file, and set all guest virtual machines' " "MTU to the same value by using a DHCP option. For information about how to " "use this option, see Configure OVS plug-" "in." msgstr "" #: ./doc/config-reference/networking/section_networking-plugins-ml2.xml48(title) msgid "Modular Layer 2 (ml2) Flat Type configuration options" msgstr "" #: ./doc/config-reference/networking/section_networking-plugins-ml2.xml53(title) msgid "Modular Layer 2 (ml2) VXLAN Type configuration options" msgstr "" #: ./doc/config-reference/networking/section_networking-plugins-ml2.xml58(title) msgid "Modular Layer 2 (ml2) Arista Mechanism configuration options" msgstr "" #: ./doc/config-reference/networking/section_networking-plugins-ml2.xml64(title) msgid "Modular Layer 2 (ml2) Cisco Mechanism configuration options" msgstr "" #: ./doc/config-reference/networking/section_networking-plugins-ml2.xml69(title) msgid "Modular Layer 2 (ml2) L2 Population Mechanism configuration options" msgstr "" #: ./doc/config-reference/networking/section_networking-plugins-ml2.xml74(title) msgid "Modular Layer 2 (ml2) Tail-f NCS Mechanism configuration options" msgstr "" #: ./doc/config-reference/networking/section_networking-options-reference.xml6(title) msgid "Networking configuration options" msgstr "" #: ./doc/config-reference/networking/section_networking-options-reference.xml7(para) msgid "" "The options and descriptions listed in this introduction are auto generated " "from the code in the Networking service project, which provides software-" "defined networking between VMs run in Compute. The list contains common " "options, while the subsections list the options for the various networking " "plug-ins." msgstr "" #: ./doc/config-reference/networking/section_networking-options-reference.xml20(para) msgid "Use the following options to alter agent-related settings." msgstr "" #: ./doc/config-reference/networking/section_networking-options-reference.xml25(title) msgid "API" msgstr "" #: ./doc/config-reference/networking/section_networking-options-reference.xml26(para) msgid "Use the following options to alter API-related settings." msgstr "" #: ./doc/config-reference/networking/section_networking-options-reference.xml31(title) msgid "Database" msgstr "" #: ./doc/config-reference/networking/section_networking-options-reference.xml32(para) msgid "Use the following options to alter Database-related settings." msgstr "" #: ./doc/config-reference/networking/section_networking-options-reference.xml37(title) msgid "Logging" msgstr "" #: ./doc/config-reference/networking/section_networking-options-reference.xml38(para) msgid "Use the following options to alter logging settings." msgstr "" #: ./doc/config-reference/networking/section_networking-options-reference.xml43(title) msgid "Metadata Agent" msgstr "" #: ./doc/config-reference/networking/section_networking-options-reference.xml44(para) msgid "" "Use the following options in the metadata_agent.ini " "file for the Metadata agent." msgstr "" #: ./doc/config-reference/networking/section_networking-options-reference.xml50(title) msgid "Policy" msgstr "" #: ./doc/config-reference/networking/section_networking-options-reference.xml51(para) msgid "" "Use the following options in the neutron.conf file to " "change policy settings." msgstr "" #: ./doc/config-reference/networking/section_networking-options-reference.xml57(title) msgid "Quotas" msgstr "" #: ./doc/config-reference/networking/section_networking-options-reference.xml58(para) msgid "" "Use the following options in the neutron.conf file for " "the quota system." msgstr "" #: ./doc/config-reference/networking/section_networking-options-reference.xml64(title) msgid "Scheduler" msgstr "" #: ./doc/config-reference/networking/section_networking-options-reference.xml65(para) msgid "" "Use the following options in the neutron.conf file to " "change scheduler settings." msgstr "" #: ./doc/config-reference/networking/section_networking-options-reference.xml71(title) msgid "Security Groups" msgstr "" #: ./doc/config-reference/networking/section_networking-options-reference.xml72(para) msgid "" "Use the following options in the configuration file for your driver to " "change security group settings." msgstr "" #: ./doc/config-reference/networking/section_networking-options-reference.xml78(title) msgid "SSL" msgstr "" #: ./doc/config-reference/networking/section_networking-options-reference.xml79(para) msgid "" "Use the following options in the neutron.conf file to " "enable SSL." msgstr "" #: ./doc/config-reference/networking/section_networking-options-reference.xml85(title) msgid "Testing" msgstr "" #: ./doc/config-reference/networking/section_networking-options-reference.xml86(para) msgid "Use the following options to alter testing-related features." msgstr "" #: ./doc/config-reference/networking/section_networking-options-reference.xml91(title) msgid "WSGI" msgstr "" #: ./doc/config-reference/networking/section_networking-options-reference.xml92(para) msgid "" "Use the following options in the neutron.conf file to " "configure the WSGI layer." msgstr "" #: ./doc/config-reference/block-storage/section_backup-drivers.xml6(title) msgid "Backup drivers" msgstr "" #: ./doc/config-reference/block-storage/section_backup-drivers.xml7(para) msgid "" "This section describes how to configure the cinder-backup service and its drivers." msgstr "" #: ./doc/config-reference/block-storage/section_backup-drivers.xml10(para) msgid "" "The volume drivers are included with the Block Storage repository (https://github.com/openstack/cinder)." " To set a backup driver, use the backup_driver flag. By " "default there is no backup driver enabled." msgstr "" #: ./doc/config-reference/block-storage/section_volume-drivers.xml6(title) msgid "Volume drivers" msgstr "" #: ./doc/config-reference/block-storage/section_volume-drivers.xml7(para) msgid "" "To use different volume drivers for the cinder-volume service, use the parameters described in these" " sections." msgstr "" #: ./doc/config-reference/block-storage/section_volume-drivers.xml10(para) msgid "" "The volume drivers are included in the Block Storage repository (https://github.com/openstack/cinder)." " To set a volume driver, use the volume_driver flag. The " "default is:" msgstr "" #: ./doc/config-reference/block-storage/section_block-storage-overview.xml6(title) msgid "Introduction to the Block Storage Service" msgstr "" #: ./doc/config-reference/block-storage/section_block-storage-overview.xml7(para) msgid "" "The Openstack Block Storage Service provides persistent block storage " "resources that OpenStack Compute instances can consume. This includes " "secondary attached storage similar to the Amazon Elastic Block Storage (EBS)" " offering. In addition, you can write images to a Block Storage device for " "Compute to use as a bootable persistent instance." msgstr "" #: ./doc/config-reference/block-storage/section_block-storage-overview.xml14(para) msgid "" "The Block Storage Service differs slightly from the Amazon EBS offering. The" " Block Storage Service does not provide a shared storage solution like NFS. " "With the Block Storage Service, you can attach a device to only one " "instance." msgstr "" #: ./doc/config-reference/block-storage/section_block-storage-overview.xml19(para) msgid "The Block Storage Service provides:" msgstr "" #: ./doc/config-reference/block-storage/section_block-storage-overview.xml22(para) msgid "" "cinder-api. A WSGI app that " "authenticates and routes requests throughout the Block Storage Service. It " "supports the OpenStack APIs only, although there is a translation that can " "be done through Compute's EC2 interface, which calls in to the cinderclient." msgstr "" #: ./doc/config-reference/block-storage/section_block-storage-overview.xml30(para) msgid "" "cinder-scheduler. Schedules and " "routes requests to the appropriate volume service. As of Grizzly; depending " "upon your configuration this may be simple round-robin scheduling to the " "running volume services, or it can be more sophisticated through the use of " "the Filter Scheduler. The Filter Scheduler is the default in Grizzly and " "enables filters on things like Capacity, Availability Zone, Volume Types, " "and Capabilities as well as custom filters." msgstr "" #: ./doc/config-reference/block-storage/section_block-storage-overview.xml38(para) msgid "" "cinder-volume. Manages Block " "Storage devices, specifically the back-end devices themselves." msgstr "" #: ./doc/config-reference/block-storage/section_block-storage-overview.xml43(para) msgid "" "cinder-backup Provides a means to" " back up a Block Storage Volume to OpenStack Object Store (SWIFT)." msgstr "" #: ./doc/config-reference/block-storage/section_block-storage-overview.xml48(para) msgid "The Block Storage Service contains the following components:" msgstr "" #: ./doc/config-reference/block-storage/section_block-storage-overview.xml52(para) msgid "" "Back-end Storage Devices. The Block " "Storage Service requires some form of back-end storage that the service is " "built on. The default implementation is to use LVM on a local volume group " "named \"cinder-volumes.\" In addition to the base driver implementation, the" " Block Storage Service also provides the means to add support for other " "storage devices to be utilized such as external Raid Arrays or other storage" " appliances. These back-end storage devices may have custom block sizes when" " using KVM or QEMU as the hypervisor." msgstr "" #: ./doc/config-reference/block-storage/section_block-storage-overview.xml65(para) msgid "" "Users and Tenants (Projects). The Block " "Storage Service is designed to be used by many different cloud computing " "consumers or customers, basically tenants on a shared system, using role-" "based access assignments. Roles control the actions that a user is allowed " "to perform. In the default configuration, most actions do not require a " "particular role, but this is configurable by the system administrator " "editing the appropriate policy.json file that maintains" " the rules. A user's access to particular volumes is limited by tenant, but " "the username and password are assigned per user. Key pairs granting access " "to a volume are enabled per user, but quotas to control resource consumption" " across available hardware resources are per tenant." msgstr "" #: ./doc/config-reference/block-storage/section_block-storage-overview.xml81(para) msgid "For tenants, quota controls are available to limit:" msgstr "" #: ./doc/config-reference/block-storage/section_block-storage-overview.xml85(para) msgid "The number of volumes that can be created" msgstr "" #: ./doc/config-reference/block-storage/section_block-storage-overview.xml89(para) msgid "The number of snapshots that can be created" msgstr "" #: ./doc/config-reference/block-storage/section_block-storage-overview.xml93(para) msgid "" "The total number of GBs allowed per tenant (shared between snapshots and " "volumes)" msgstr "" #: ./doc/config-reference/block-storage/section_block-storage-overview.xml97(para) msgid "" "You can revise the default quota values with the cinder CLI, so the limits " "placed by quotas are editable by admin users." msgstr "" #: ./doc/config-reference/block-storage/section_block-storage-overview.xml100(para) msgid "" "Volumes, Snapshots, and Backups. The " "basic resources offered by the Block Storage Service are volumes and " "snapshots, which are derived from volumes, and backups:" msgstr "" #: ./doc/config-reference/block-storage/section_block-storage-overview.xml107(para) msgid "" "Volumes. Allocated block storage " "resources that can be attached to instances as secondary storage or they can" " be used as the root store to boot instances. Volumes are persistent R/W " "block storage devices most commonly attached to the Compute node through " "iSCSI." msgstr "" #: ./doc/config-reference/block-storage/section_block-storage-overview.xml116(para) msgid "" "Snapshots. A read-only point in time copy" " of a volume. The snapshot can be created from a volume that is currently in" " use (through the use of '--force True') or in an available state. The " "snapshot can then be used to create a new volume through create from " "snapshot." msgstr "" #: ./doc/config-reference/block-storage/section_block-storage-overview.xml125(para) msgid "" "Backups. An archived copy of a volume " "currently stored in OpenStack Object Storage (Swift)." msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml6(title) msgid "VMware VMDK driver" msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml7(para) msgid "" "Use the VMware VMDK driver to enable management of the OpenStack Block " "Storage volumes on vCenter-managed data stores. Volumes are backed by VMDK " "files on data stores using any VMware-compatible storage technology such as " "NFS, iSCSI, FiberChannel, and vSAN." msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml13(title) #: ./doc/config-reference/block-storage/drivers/xen-sm-driver.xml134(title) #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml72(title) msgid "Configuration" msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml14(para) msgid "" "The recommended volume driver for OpenStack Block Storage is the VMware " "vCenter VMDK driver. When you configure the driver, you must match it with " "the appropriate OpenStack Compute driver from VMware and both drivers must " "point to the same server." msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml19(para) msgid "" "For example, in the nova.conf file, use this option to " "define the Compute driver:" msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml22(para) msgid "" "In the cinder.conf file, use this option to define the " "volume driver:" msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml25(para) msgid "" "The following table lists various options that the drivers support for the " "OpenStack Block Storage configuration (cinder.conf):" msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml32(para) msgid "" "The VMware VMDK drivers support the creation of VMDK disk files of type " "thin, thick, or " "eagerZeroedThick. Use the vmware:vmdk_type " "extra spec key with the appropriate value to specify the VMDK disk file " "type. The following table captures the mapping between the extra spec entry " "and the VMDK disk file type:" msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml40(caption) msgid "Extra spec entry to VMDK disk file type mapping" msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml44(td) msgid "Disk file type" msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml45(td) #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml94(td) msgid "Extra spec key" msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml46(td) #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml95(td) msgid "Extra spec value" msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml52(td) #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml57(td) #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml62(td) msgid "vmware:vmdk_type" msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml56(td) #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml58(td) msgid "thick" msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml61(td) #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml63(td) msgid "eagerZeroedThick" msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml67(para) msgid "" "If no vmdk_type extra spec entry is specified, the default disk" " file type is thin." msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml70(para) msgid "" "The example below shows how to create a thick VMDK volume using" " the appropriate vmdk_type:" msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml80(title) #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml93(td) msgid "Clone type" msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml81(para) msgid "" "With the VMware VMDK drivers, you can create a volume from another source " "volume or from a snapshot point. The VMware vCenter VMDK driver supports " "clone types full and linked/fast. The " "clone type is specified using the vmware:clone_type extra spec " "key with the appropriate value. The following table captures the mapping for" " clone types:" msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml90(caption) msgid "Extra spec entry to clone type mapping" msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml100(td) #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml102(td) msgid "full" msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml101(td) #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml106(td) msgid "vmware:clone_type" msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml105(td) msgid "linked/fast" msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml107(td) msgid "linked" msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml111(para) msgid "If not specified, the default clone type is full." msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml113(para) msgid "" "The following is an example of linked cloning from another source volume:" msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml120(para) msgid "" "Note: The VMware ESX VMDK driver ignores the extra spec entry and always " "creates a full clone." msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml125(title) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml13(title) #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml31(title) #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml42(title) msgid "Supported operations" msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml126(para) msgid "" "The following operations are supported by the VMware vCenter and ESX VMDK " "drivers:" msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml130(para) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml17(para) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml48(para) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml74(para) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml93(para) #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml36(para) #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml30(para) msgid "Create volume" msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml133(para) msgid "" "Create volume from another source volume. (Supported only if source volume " "is not attached to an instance.)" msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml138(para) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml35(para) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml111(para) #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml66(para) #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml60(para) msgid "Create volume from snapshot" msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml141(para) msgid "Create volume from glance image" msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml144(para) msgid "" "Attach volume (When a volume is attached to an instance, a reconfigure " "operation is performed on the instance to add the volume's VMDK to it. The " "user must manually rescan and mount the device from within the guest " "operating system.)" msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml151(para) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml26(para) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml57(para) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml83(para) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml102(para) #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml45(para) #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml39(para) msgid "Detach volume" msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml154(para) msgid "" "Create snapshot (Allowed only if volume is not attached to an instance.)" msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml158(para) msgid "" "Delete snapshot (Allowed only if volume is not attached to an instance.)" msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml162(para) msgid "" "Upload as image to glance (Allowed only if volume is not attached to an " "instance.)" msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml167(para) msgid "" "Although the VMware ESX VMDK driver supports these operations, it has not " "been extensively tested." msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml172(title) msgid "Data store selection" msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml173(para) msgid "" "When creating a volume, the driver chooses a data store that has sufficient " "free space and has the highest freespace/totalspace " "metric value." msgstr "" #: ./doc/config-reference/block-storage/drivers/vmware-vmdk-driver.xml177(para) msgid "" "When a volume is attached to an instance, the driver attempts to place the " "volume under the instance's ESX host on a data store that is selected using " "the strategy above." msgstr "" #: ./doc/config-reference/block-storage/drivers/nexenta-volume-driver.xml4(title) msgid "Nexenta drivers" msgstr "" #: ./doc/config-reference/block-storage/drivers/nexenta-volume-driver.xml5(para) msgid "" "NexentaStor Appliance is NAS/SAN software platform designed for building " "reliable and fast network storage arrays. The Nexenta Storage Appliance uses" " ZFS as a disk management system. NexentaStor can serve as a storage node " "for the OpenStack and for the virtual servers through iSCSI and NFS " "protocols." msgstr "" #: ./doc/config-reference/block-storage/drivers/nexenta-volume-driver.xml11(para) msgid "" "With the NFS option, every Compute volume is represented by a directory " "designated to be its own file system in the ZFS file system. These file " "systems are exported using NFS." msgstr "" #: ./doc/config-reference/block-storage/drivers/nexenta-volume-driver.xml14(para) msgid "" "With either option some minimal setup is required to tell OpenStack which " "NexentaStor servers are being used, whether they are supporting iSCSI and/or" " NFS and how to access each of the servers." msgstr "" #: ./doc/config-reference/block-storage/drivers/nexenta-volume-driver.xml18(para) msgid "" "Typically the only operation required on the NexentaStor servers is to " "create the containing directory for the iSCSI or NFS exports. For NFS this " "containing directory must be explicitly exported via NFS. There is no " "software that must be installed on the NexentaStor servers; they are " "controlled using existing management plane interfaces." msgstr "" #: ./doc/config-reference/block-storage/drivers/nexenta-volume-driver.xml26(title) msgid "Nexenta iSCSI driver" msgstr "" #: ./doc/config-reference/block-storage/drivers/nexenta-volume-driver.xml27(para) msgid "" "The Nexenta iSCSI driver allows you to use NexentaStor appliance to store " "Compute volumes. Every Compute volume is represented by a single zvol in a " "predefined Nexenta namespace. For every new volume the driver creates a " "iSCSI target and iSCSI target group that are used to access it from compute " "hosts." msgstr "" #: ./doc/config-reference/block-storage/drivers/nexenta-volume-driver.xml33(para) msgid "" "The Nexenta iSCSI volume driver should work with all versions of " "NexentaStor. The NexentaStor appliance must be installed and configured " "according to the relevant Nexenta documentation. A pool and an enclosing " "namespace must be created for all iSCSI volumes to be accessed through the " "volume driver. This should be done as specified in the release specific " "NexentaStor documentation." msgstr "" #: ./doc/config-reference/block-storage/drivers/nexenta-volume-driver.xml40(para) msgid "" "The NexentaStor Appliance iSCSI driver is selected using the normal " "procedures for one or multiple back-end volume drivers. You must configure " "these items for each NexentaStor appliance that the iSCSI volume driver " "controls:" msgstr "" #: ./doc/config-reference/block-storage/drivers/nexenta-volume-driver.xml46(title) msgid "Enable the Nexenta iSCSI driver and related options" msgstr "" #: ./doc/config-reference/block-storage/drivers/nexenta-volume-driver.xml48(para) msgid "This table contains the options supported by the Nexenta iSCSI driver." msgstr "" #: ./doc/config-reference/block-storage/drivers/nexenta-volume-driver.xml52(para) msgid "" "To use Compute with the Nexenta iSCSI driver, first set the " "volume_driver:" msgstr "" #: ./doc/config-reference/block-storage/drivers/nexenta-volume-driver.xml56(para) msgid "" "Then, set the nexenta_host parameter and other parameters from " "the table, if needed." msgstr "" #: ./doc/config-reference/block-storage/drivers/nexenta-volume-driver.xml63(title) msgid "Nexenta NFS driver" msgstr "" #: ./doc/config-reference/block-storage/drivers/nexenta-volume-driver.xml64(para) msgid "" "The Nexenta NFS driver allows you to use NexentaStor appliance to store " "Compute volumes via NFS. Every Compute volume is represented by a single NFS" " file within a shared directory." msgstr "" #: ./doc/config-reference/block-storage/drivers/nexenta-volume-driver.xml68(para) msgid "" "While the NFS protocols standardize file access for users, they do not " "standardize administrative actions such as taking snapshots or replicating " "file systems. The Openstack Volume Drivers bring a common interface to these" " operations. The Nexenta NFS driver implements these standard actions using " "the ZFS management plane that already is deployed on NexentaStor appliances." msgstr "" #: ./doc/config-reference/block-storage/drivers/nexenta-volume-driver.xml75(para) msgid "" "The Nexenta NFS volume driver should work with all versions of NexentaStor. " "The NexentaStor appliance must be installed and configured according to the " "relevant Nexenta documentation. A single parent file system must be created " "for all virtual disk directories supported for OpenStack. This directory " "must be created and exported on each NexentaStor appliance. This should be " "done as specified in the release specific NexentaStor documentation." msgstr "" #: ./doc/config-reference/block-storage/drivers/nexenta-volume-driver.xml84(title) msgid "Enable the Nexenta NFS driver and related options" msgstr "" #: ./doc/config-reference/block-storage/drivers/nexenta-volume-driver.xml86(para) msgid "" "To use Compute with the Nexenta NFS driver, first set the " "volume_driver:" msgstr "" #: ./doc/config-reference/block-storage/drivers/nexenta-volume-driver.xml91(para) msgid "" "The following table contains the options supported by the Nexenta NFS " "driver." msgstr "" #: ./doc/config-reference/block-storage/drivers/nexenta-volume-driver.xml95(para) msgid "" "Add your list of Nexenta NFS servers to the file you specified with the " "nexenta_shares_config option. For example, if the value of this" " option was set to /etc/cinder/nfs_shares, then:" msgstr "" #: ./doc/config-reference/block-storage/drivers/nexenta-volume-driver.xml105(para) msgid "Comments are allowed in this file. They begin with a #." msgstr "" #: ./doc/config-reference/block-storage/drivers/nexenta-volume-driver.xml107(para) msgid "" "Each line in this file represents a NFS share. The first part of the line is" " the NFS share URL, the second is the connection URL to the NexentaStor " "Appliance." msgstr "" #: ./doc/config-reference/block-storage/drivers/zadara-volume-driver.xml6(title) msgid "Zadara" msgstr "" #: ./doc/config-reference/block-storage/drivers/zadara-volume-driver.xml7(para) msgid "" "There is a volume back-end for Zadara. Set the following in your " "cinder.conf, and use the following options to configure" " it." msgstr "" #. When image changes, this message will be marked fuzzy or untranslated for #. you. #. It doesn't matter what you translate it to: it's not used at all. #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml286(None) msgid "" "@@image: " "'../../../common/figures/coraid/Repository_Creation_Plan_screen.png'; " "md5=83038804978648c2db4001a46c11f8ba" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml6(title) msgid "Coraid AoE driver configuration" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml7(para) msgid "" "Coraid storage appliances can provide block-level storage to OpenStack " "instances. Coraid storage appliances use the low-latency ATA-over-Ethernet " "(ATA) protocol to provide high-bandwidth data transfer between hosts and " "data on the network." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml11(para) msgid "Once configured for OpenStack, you can:" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml14(para) msgid "Create, delete, attach, and detach block storage volumes." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml18(para) msgid "Create, list, and delete volume snapshots." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml21(para) msgid "" "Create a volume from a snapshot, copy an image to a volume, copy a volume to" " an image, clone a volume, and get volume statistics." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml26(para) msgid "" "This document describes how to configure the OpenStack Block Storage Service" " for use with Coraid storage appliances." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml29(title) msgid "Terminology" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml30(para) msgid "These terms are used in this section:" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml34(th) msgid "Term" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml35(th) msgid "Definition" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml40(td) msgid "AoE" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml41(td) msgid "ATA-over-Ethernet protocol" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml44(td) msgid "EtherCloud Storage Manager (ESM)" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml45(td) msgid "" "ESM provides live monitoring and management of EtherDrive appliances that " "use the AoE protocol, such as the SRX and VSX." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml50(td) msgid "Fully-Qualified Repository Name (FQRN)" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml53(replaceable) msgid "performance_class" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml53(replaceable) msgid "availability_class" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml53(replaceable) msgid "profile_name" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml53(replaceable) msgid "repository_name" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml51(td) msgid "" "The FQRN is the full identifier of a storage profile. FQRN syntax is: " "" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml56(td) msgid "SAN" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml57(td) msgid "Storage Area Network" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml60(td) msgid "SRX" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml61(td) msgid "Coraid EtherDrive SRX block storage appliance" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml64(td) msgid "VSX" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml65(td) msgid "Coraid EtherDrive VSX storage virtualization appliance" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml72(title) #: ./doc/config-reference/block-storage/drivers/xenapi-nfs.xml14(title) msgid "Requirements" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml73(para) msgid "" "To support the OpenStack Block Storage Service, your SAN must include an SRX" " for physical storage, a VSX running at least CorOS v2.0.6 for snapshot " "support, and an ESM running at least v2.1.1 for storage repository " "orchestration. Ensure that all storage appliances are installed and " "connected to your network before you configure OpenStack volumes." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml79(para) msgid "" "So that the node can communicate with the SAN, you must install the Coraid " "AoE Linux driver on each compute node on the network that runs an OpenStack " "instance." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml84(title) msgid "Overview" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml85(para) msgid "" "To configure the OpenStack Block Storage for use with Coraid storage " "appliances, perform the following procedures:" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml89(para) msgid "" "Download and install the " "Coraid Linux AoE driver." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml93(para) msgid "" "Create a storage profile " "by using the Coraid ESM GUI." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml97(para) msgid "" "Create a storage " "repository by using the ESM GUI and record the FQRN." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml102(para) msgid "" "Configure the " "cinder.conf file." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml106(para) msgid "" "Create and " "associate a block storage volume type." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml113(title) msgid "Install the Coraid AoE driver" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml114(para) msgid "" "Install the Coraid AoE driver on every compute node that will require access" " to block storage." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml116(para) msgid "" "The latest AoE drivers will always be located at http://support.coraid.com/support/linux/." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml119(para) msgid "" "To download and install the AoE driver, follow the instructions below, " "replacing “aoeXXX” with the AoE driver file name:" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml124(para) msgid "Download the latest Coraid AoE driver." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml130(para) msgid "Unpack the AoE driver." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml133(para) msgid "Install the AoE driver." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml139(para) msgid "Initialize the AoE driver." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml145(para) msgid "" "Optionally, specify the Ethernet interfaces that the node can use to " "communicate with the SAN." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml147(para) msgid "" "The AoE driver may use every Ethernet interface available to the node unless" " limited with the aoe_iflist parameter. For more " "information about the aoe_iflist parameter, see the " "aoe readme file included with the AoE driver." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml154(replaceable) msgid "eth1 eth2 ..." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml160(title) msgid "Create a storage profile" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml161(para) msgid "To create a storage profile using the ESM GUI:" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml164(para) msgid "Log in to the ESM." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml167(para) msgid "" "Click Storage Profiles in the SAN " "Domain pane." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml171(para) msgid "" "Choose Menu > Create Storage Profile. If the " "option is unavailable, you might not have appropriate permissions. Make sure" " you are logged in to the ESM as the SAN administrator." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml177(para) msgid "Use the storage class selector to select a storage class." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml179(para) msgid "" "Each storage class includes performance and availability criteria (see the " "Storage Classes topic in the ESM Online Help for information on the " "different options)." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml184(para) msgid "" "Select a RAID type (if more than one is available) for the selected profile " "type." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml188(para) msgid "Type a Storage Profile name." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml189(para) msgid "" "The name is restricted to alphanumeric characters, underscore (_), and " "hyphen (-), and cannot exceed 32 characters." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml194(para) msgid "Select the drive size from the drop-down menu." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml197(para) msgid "" "Select the number of drives to be initialized for each RAID (LUN) from the " "drop-down menu (if the selected RAID type requires multiple drives)." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml202(para) msgid "" "Type the number of RAID sets (LUNs) you want to create in the repository by " "using this profile." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml206(para) msgid "Click Next." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml211(title) msgid "Create a storage repository and get the FQRN" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml212(para) msgid "" "Create a storage repository and get its fully qualified repository name " "(FQRN):" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml216(para) msgid "Access the Create Storage Repository dialog box." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml220(para) msgid "Type a Storage Repository name." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml221(para) msgid "" "The name is restricted to alphanumeric characters, underscore (_), hyphen " "(-), and cannot exceed 32 characters." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml226(para) msgid "" "Click Limited or Unlimited to " "indicate the maximum repository size." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml229(para) msgid "" "Limited sets the amount of space that can be " "allocated to the repository. Specify the size in TB, GB, or MB." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml232(para) msgid "" "When the difference between the reserved space and the space already " "allocated to LUNs is less than is required by a LUN allocation request, the " "reserved space is increased until the repository limit is reached." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml237(para) msgid "" "The reserved space does not include space used for parity or space used for " "mirrors. If parity and/or mirrors are required, the actual space allocated " "to the repository from the SAN is greater than that specified in reserved " "space." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml243(para) msgid "" "Unlimited—Unlimited means that the amount" " of space allocated to the repository is unlimited and additional space is " "allocated to the repository automatically when space is required and " "available." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml249(para) msgid "" "Drives specified in the associated Storage Profile must be available on the " "SAN in order to allocate additional resources." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml255(para) msgid "Check the Resizeable LUN box." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml257(para) msgid "This is required for OpenStack volumes." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml259(para) msgid "" "If the Storage Profile associated with the repository has platinum " "availability, the Resizeable LUN box is automatically checked." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml265(para) msgid "" "Check the Show Allocation Plan API calls box. Click " "Next." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml270(para) msgid "Record the FQRN and click Finish." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml272(para) msgid "" "The FQRN is located in the first line of output following the " "Plan keyword in the Repository Creation " "Plan window. The FQRN syntax is " "performance_classavailability_classprofile_namerepository_name." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml277(para) msgid "" "In this example, the FQRN is Bronze-" "Platinum:BP1000:OSTest, and is highlighted." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml281(title) msgid "Repository Creation Plan screen" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml290(para) msgid "" "Record the FQRN; it is a required parameter later in the configuration " "procedure." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml296(title) msgid "Configure options in the cinder.conf file" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml297(para) msgid "" "Edit or add the following lines to the file " "/etc/cinder/cinder.conf:" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml300(replaceable) msgid "ESM_IP_address" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml301(replaceable) #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml340(option) msgid "username" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml302(replaceable) msgid "Access_Control_Group_name" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml303(replaceable) #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml350(option) msgid "password" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml304(replaceable) #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml352(replaceable) #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml370(replaceable) msgid "coraid_repository_key" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml306(para) msgid "" "Access to storage devices and storage repositories can be controlled using " "Access Control Groups configured in ESM. Configuring " "cinder.conf to log on to ESM as the SAN administrator " "(user name admin), will grant full access to the devices " "and repositories configured in ESM." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml312(para) msgid "" "Optionally, you can configure an ESM Access Control Group and user. Then, " "use the cinder.conf file to configure access to the ESM" " through that group, and user limits access from the OpenStack instance to " "devices and storage repositories that are defined in the group." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml317(para) msgid "" "To manage access to the SAN by using Access Control Groups, you must enable " "the Use Access Control setting in the ESM System " "Setup > Security screen." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml321(para) msgid "For more information, see the ESM Online Help." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml324(title) msgid "Create and associate a volume type" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml325(para) msgid "Create and associate a volume with the ESM storage repository." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml329(para) msgid "Restart Cinder." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml335(para) msgid "Create a volume." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml336(replaceable) msgid "volume_type_name" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml337(para) msgid "" "where volume_type_name is the name you assign the" " volume. You will see output similar to the following:" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml345(para) msgid "Record the value in the ID field; you use this value in the next step." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml349(para) msgid "Associate the volume type with the Storage Repository." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml352(replaceable) #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml363(replaceable) msgid "UUID" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml352(replaceable) #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml378(replaceable) msgid "FQRN" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml357(th) msgid "Variable" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml358(th) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml311(td) #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml202(td) #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml319(td) #: ./doc/config-reference/block-storage/drivers/ibm-gpfs-volume-driver.xml80(th) msgid "Description" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml364(td) msgid "" "The ID returned from the command. You can use the " " command to recover the ID." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml373(filename) msgid "cinder.conf" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml375(literal) msgid "coraid_repository" msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml371(td) msgid "" "The key name used to associate the Cinder volume type with the ESM in the " " file. If no key name was defined, this is default value for" " ." msgstr "" #: ./doc/config-reference/block-storage/drivers/coraid-driver.xml379(td) msgid "The FQRN recorded during the Create Storage Repository process." msgstr "" #: ./doc/config-reference/block-storage/drivers/solidfire-volume-driver.xml5(title) msgid "SolidFire" msgstr "" #: ./doc/config-reference/block-storage/drivers/solidfire-volume-driver.xml6(para) msgid "" "The SolidFire Cluster is a high performance all SSD iSCSI storage device " "that provides massive scale out capability and extreme fault tolerance. A " "key feature of the SolidFire cluster is the ability to set and modify during" " operation specific QoS levels on a volume for volume basis. The SolidFire " "cluster offers this along with de-duplication, compression, and an " "architecture that takes full advantage of SSDs." msgstr "" #: ./doc/config-reference/block-storage/drivers/solidfire-volume-driver.xml14(para) msgid "" "To configure the use of a SolidFire cluster with Block Storage, modify your " "cinder.conf file as follows:" msgstr "" #: ./doc/config-reference/block-storage/drivers/solidfire-volume-driver.xml23(para) msgid "" "The SolidFire driver creates a unique account prefixed with $cinder-volume-service-hostname-$tenant-id on the SolidFire " "cluster for each tenant that accesses the cluster through the Volume API. " "Unfortunately, this account formation results in issues for High " "Availability (HA) installations and installations where the cinder-volume service can move to a new node." " HA installations can return an Account Not Found " "error because the call to the SolidFire cluster is not always going to be " "sent from the same node. In installations where the cinder-volume service moves to a new node, " "the same issue can occur when you perform operations on existing volumes, " "such as clone, extend, delete, and so on." msgstr "" #: ./doc/config-reference/block-storage/drivers/solidfire-volume-driver.xml41(para) msgid "" "Set the sf_account_prefix option to an empty string ('') " "in the cinder.conf file. This setting results in unique" " accounts being created on the SolidFire cluster, but the accounts are " "prefixed with the tenant-id or any unique identifier that you choose and are" " independent of the host where the cinder-" "volume service resides." msgstr "" #: ./doc/config-reference/block-storage/drivers/xen-sm-driver.xml4(title) msgid "XenAPI Storage Manager volume driver" msgstr "" #: ./doc/config-reference/block-storage/drivers/xen-sm-driver.xml5(para) msgid "" "The Xen Storage Manager volume driver (xensm) is a XenAPI hypervisor " "specific volume driver, and can be used to provide basic storage " "functionality, including volume creation and destruction, on a number of " "different storage back-ends. It also enables the capability of using more " "sophisticated storage back-ends for operations like cloning/snapshots, and " "so on. Some of the storage plug-ins that are already supported in Citrix " "XenServer and Xen Cloud Platform (XCP) are:" msgstr "" #: ./doc/config-reference/block-storage/drivers/xen-sm-driver.xml15(para) msgid "" "NFS VHD: Storage repository (SR) plug-in that stores disks as Virtual Hard " "Disk (VHD) files on a remote Network File System (NFS)." msgstr "" #: ./doc/config-reference/block-storage/drivers/xen-sm-driver.xml20(para) msgid "" "Local VHD on LVM: SR plug-in that represents disks as VHD disks on Logical " "Volumes (LVM) within a locally-attached Volume Group." msgstr "" #: ./doc/config-reference/block-storage/drivers/xen-sm-driver.xml25(para) msgid "" "HBA LUN-per-VDI driver: SR plug-in that represents Logical Units (LUs) as " "Virtual Disk Images (VDIs) sourced by host bus adapters (HBAs). For example," " hardware-based iSCSI or FC support." msgstr "" #: ./doc/config-reference/block-storage/drivers/xen-sm-driver.xml31(para) msgid "" "NetApp: SR driver for mapping of LUNs to VDIs on a NETAPP server, providing " "use of fast snapshot and clone features on the filer." msgstr "" #: ./doc/config-reference/block-storage/drivers/xen-sm-driver.xml36(para) msgid "" "LVHD over FC: SR plug-in that represents disks as VHDs on Logical Volumes " "within a Volume Group created on an HBA LUN. For example, hardware-based " "iSCSI or FC support." msgstr "" #: ./doc/config-reference/block-storage/drivers/xen-sm-driver.xml42(para) msgid "" "iSCSI: Base ISCSI SR driver, provides a LUN-per-VDI. Does not support " "creation of VDIs but accesses existing LUNs on a target." msgstr "" #: ./doc/config-reference/block-storage/drivers/xen-sm-driver.xml47(para) msgid "" "LVHD over iSCSI: SR plug-in that represents disks as Logical Volumes within " "a Volume Group created on an iSCSI LUN." msgstr "" #: ./doc/config-reference/block-storage/drivers/xen-sm-driver.xml52(para) msgid "" "EqualLogic: SR driver for mapping of LUNs to VDIs on a EQUALLOGIC array " "group, providing use of fast snapshot and clone features on the array." msgstr "" #: ./doc/config-reference/block-storage/drivers/xen-sm-driver.xml58(title) msgid "Design and operation" msgstr "" #: ./doc/config-reference/block-storage/drivers/xen-sm-driver.xml60(title) msgid "Definitions" msgstr "" #: ./doc/config-reference/block-storage/drivers/xen-sm-driver.xml63(para) msgid "" "Back-end: A term for a particular storage" " back-end. This could be iSCSI, NFS, NetApp, and so on." msgstr "" #: ./doc/config-reference/block-storage/drivers/xen-sm-driver.xml68(para) msgid "" "Back-end-config: All the parameters " "required to connect to a specific back-end. For example, for NFS, this would" " be the server, path, and so on." msgstr "" #: ./doc/config-reference/block-storage/drivers/xen-sm-driver.xml75(para) msgid "" "Flavor: This term is equivalent to volume" " \"types\". A user friendly term to specify some notion of quality of " "service. For example, \"gold\" might mean that the volumes use a back-end " "where backups are possible. A flavor can be associated with multiple back-" "ends. The volume scheduler, with the help of the driver, decides which back-" "end is used to create a volume of a particular flavor. Currently, the driver" " uses a simple \"first-fit\" policy, where the first back-end that can " "successfully create this volume is the one that is used." msgstr "" #: ./doc/config-reference/block-storage/drivers/xen-sm-driver.xml93(title) msgid "Operation" msgstr "" #: ./doc/config-reference/block-storage/drivers/xen-sm-driver.xml94(para) msgid "" "The admin uses the nova-manage command detailed below to add flavors and " "back-ends." msgstr "" #: ./doc/config-reference/block-storage/drivers/xen-sm-driver.xml96(para) msgid "" "One or more cinder-volume service" " instances are deployed for each availability zone. When an instance is " "started, it creates storage repositories (SRs) to connect to the back-ends " "available within that zone. All cinder-" "volume instances within a zone can see all the available back-" "ends. These instances are completely symmetric and hence should be able to " "service any create_volume request within the zone." msgstr "" #: ./doc/config-reference/block-storage/drivers/xen-sm-driver.xml108(title) msgid "On XenServer, PV guests required" msgstr "" #: ./doc/config-reference/block-storage/drivers/xen-sm-driver.xml109(para) msgid "" "Note that when using XenServer you can only attach a volume to a PV guest." msgstr "" #: ./doc/config-reference/block-storage/drivers/xen-sm-driver.xml115(title) msgid "Configure XenAPI Storage Manager" msgstr "" #: ./doc/config-reference/block-storage/drivers/xen-sm-driver.xml120(para) msgid "" "xensm requires that you use either Citrix XenServer or XCP as the " "hypervisor. The NetApp and EqualLogic back-ends are not supported on XCP." msgstr "" #: ./doc/config-reference/block-storage/drivers/xen-sm-driver.xml126(para) msgid "" "Ensure all hosts running volume and " "compute services have connectivity to the storage system." msgstr "" #: ./doc/config-reference/block-storage/drivers/xen-sm-driver.xml141(systemitem) #: ./doc/config-reference/block-storage/drivers/xen-sm-driver.xml181(systemitem) msgid "nova-compute" msgstr "" #: ./doc/config-reference/block-storage/drivers/xen-sm-driver.xml138(emphasis) msgid "" "Set the following configuration options for the nova volume service: " "( also requires the volume_driver configuration option.)" msgstr "" #: ./doc/config-reference/block-storage/drivers/xen-sm-driver.xml152(emphasis) msgid "" "You must create the back-end configurations that the volume driver uses " "before you start the volume service." msgstr "" #: ./doc/config-reference/block-storage/drivers/xen-sm-driver.xml161(para) msgid "" "SR type and configuration connection parameters are in keeping with the " "XenAPI Command " "Line Interface." msgstr "" #: ./doc/config-reference/block-storage/drivers/xen-sm-driver.xml168(para) msgid "Example: For the NFS storage manager plug-in, run these commands:" msgstr "" #: ./doc/config-reference/block-storage/drivers/xen-sm-driver.xml179(systemitem) msgid "cinder-volume" msgstr "" #: ./doc/config-reference/block-storage/drivers/xen-sm-driver.xml177(emphasis) msgid "" "Start and with the new configuration " "options." msgstr "" #: ./doc/config-reference/block-storage/drivers/xen-sm-driver.xml188(title) msgid "Create and access the volumes from VMs" msgstr "" #: ./doc/config-reference/block-storage/drivers/xen-sm-driver.xml189(para) msgid "" "Currently, the flavors have not been tied to the volume types API. As a " "result, we simply end up creating volumes in a \"first fit\" order on the " "given back-ends." msgstr "" #: ./doc/config-reference/block-storage/drivers/xen-sm-driver.xml193(para) msgid "" "Use the standard or OpenStack API commands (such as volume " "extensions) to create, destroy, attach, or detach volumes." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-xiv-volume-driver.xml5(title) msgid "IBM XIV/DS8K volume driver" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-xiv-volume-driver.xml6(para) msgid "" "There is a unified volume back-end for IBM XIV and DS8K storage. Set the " "following in your cinder.conf, and use the following " "options to configure it." msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml5(title) msgid "NetApp unified driver" msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml6(para) msgid "" "The NetApp® unified driver is a block storage driver that supports multiple " "storage families and protocols. A storage family corresponds to storage " "systems built on different NetApp technologies such as clustered Data ONTAP®" " and Data ONTAP operating in 7-Mode. The storage protocol refers to the " "protocol used to initiate data storage and access operations on those " "storage systems like iSCSI and NFS. The NetApp unified driver can be " "configured to provision and manage OpenStack volumes on a given storage " "family using a specified storage protocol. The OpenStack volumes can then be" " used for accessing and storing data using the storage protocol on the " "storage family system. The NetApp unified driver is an extensible interface " "that can support new storage families and protocols." msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml21(title) msgid "NetApp clustered Data ONTAP storage family" msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml22(para) msgid "" "The NetApp clustered Data ONTAP storage family represents a configuration " "group which provides OpenStack compute instances access to clustered Data " "ONTAP storage systems. At present it can be configured in Cinder to work " "with iSCSI and NFS storage protocols." msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml28(title) msgid "NetApp iSCSI configuration for clustered Data ONTAP" msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml30(para) msgid "" "The NetApp iSCSI configuration for clustered Data ONTAP is an interface from" " OpenStack to clustered Data ONTAP storage systems for provisioning and " "managing the SAN block storage entity; that is, a NetApp LUN which can be " "accessed using the iSCSI protocol." msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml35(para) msgid "" "The iSCSI configuration for clustered Data ONTAP is a direct interface from " "Cinder to the clustered Data ONTAP instance and as such does not require " "additional management software to achieve the desired functionality. It uses" " NetApp APIs to interact with the clustered Data ONTAP instance." msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml42(title) msgid "" "Configuration options for clustered Data ONTAP family with iSCSI protocol" msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml44(para) msgid "" "Configure the volume driver, storage family and storage protocol to the " "NetApp unified driver, clustered Data ONTAP, and iSCSI respectively by " "setting the volume_driver, " "netapp_storage_family and " "netapp_storage_protocol options in " "cinder.conf as follows:" msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml62(para) #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml194(para) msgid "" "You must override the default value of " "netapp_storage_protocol with iscsi in " "order to utilize the iSCSI protocol." msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml67(para) #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml115(para) msgid "" "If you specify an account in the netapp_login that only " "has virtual storage server (Vserver) administration privileges (rather than " "cluster-wide administration privileges), some advanced features of the " "NetApp unified driver will not work and you may see warnings in the Cinder " "logs." msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml74(para) msgid "" "For more information on these options and other deployment and operational " "scenarios, visit the OpenStack NetApp " "community." msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml81(title) msgid "NetApp NFS configuration for clustered Data ONTAP" msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml83(para) msgid "" "The NetApp NFS configuration for clustered Data ONTAP is an interface from " "OpenStack to a clustered Data ONTAP system for provisioning and managing " "OpenStack volumes on NFS exports provided by the clustered Data ONTAP system" " that are accessed using the NFS protocol." msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml88(para) msgid "" "The NFS configuration for clustered Data ONTAP is a direct interface from " "Cinder to the clustered Data ONTAP instance and as such does not require any" " additional management software to achieve the desired functionality. It " "uses NetApp APIs to interact with the clustered Data ONTAP instance." msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml94(title) msgid "" "Configuration options for the clustered Data ONTAP family with NFS protocol" msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml96(para) msgid "" "Configure the volume driver, storage family and storage protocol to NetApp " "unified driver, clustered Data ONTAP, and NFS respectively by setting the " "volume_driver, netapp_storage_family " "and netapp_storage_protocol options in " "cinder.conf as follows:" msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml122(para) #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml199(para) #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml239(para) msgid "" "For more information on these options and other deployment and operational " "scenarios, visit the OpenStack NetApp " "community." msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml126(title) msgid "NetApp-supported extra specs for clustered Data ONTAP" msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml128(para) msgid "" "Extra specs allow individual vendors to specify additional filter criteria " "which the Cinder scheduler can use when evaluating which volume node should " "fulfill a volume provisioning request. When using the NetApp unified driver " "with a clustered Data ONTAP storage system, you can leverage extra specs " "with Cinder volume types to ensure that Cinder volumes are created on " "storage backends that have certain properties (e.g. QoS, mirroring, " "compression) configured." msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml135(para) msgid "" "Extra specs are associated with Cinder volume types, so that when users " "request volumes of a particular volume type, they will be created on storage" " backends that meet the list of requirements (e.g. available space, extra " "specs, etc). You can use the specs in the table later in this section when " "defining Cinder volume types using the cinder type-key " "command." msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml142(para) msgid "" "It is recommended to only set the value of extra specs to " "True when combining multiple specs to enforce a certain " "logic set. If you desire to remove volumes with a certain feature enabled " "from consideration from the Cinder volume scheduler, be sure to use the " "negated spec name with a value of True rather than " "setting the positive spec to a value of False." msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml153(title) msgid "NetApp Data ONTAP operating in 7-Mode storage family" msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml154(para) msgid "" "The NetApp Data ONTAP operating in 7-Mode storage family represents a " "configuration group which provides OpenStack compute instances access to " "7-Mode storage systems. At present it can be configured in Cinder to work " "with iSCSI and NFS storage protocols." msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml160(title) msgid "NetApp iSCSI configuration for Data ONTAP operating in 7-Mode" msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml162(para) msgid "" "The NetApp iSCSI configuration for Data ONTAP operating in 7-Mode is an " "interface from OpenStack to Data ONTAP operating in 7-Mode storage systems " "for provisioning and managing the SAN block storage entity, that is, a LUN " "which can be accessed using iSCSI protocol." msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml167(para) msgid "" "The iSCSI configuration for Data ONTAP operating in 7-Mode is a direct " "interface from OpenStack to Data ONTAP operating in 7-Mode storage system " "and it does not require additional management software to achieve the " "desired functionality. It uses NetApp ONTAPI to interact with the Data ONTAP" " operating in 7-Mode storage system." msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml174(title) msgid "" "Configuration options for the Data ONTAP operating in 7-Mode storage family " "with iSCSI protocol" msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml176(para) msgid "" "Configure the volume driver, storage family and storage protocol to the " "NetApp unified driver, Data ONTAP operating in 7-Mode, and iSCSI " "respectively by setting the volume_driver, " "netapp_storage_family and " "netapp_storage_protocol options in " "cinder.conf as follows:" msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml203(title) msgid "NetApp NFS configuration for Data ONTAP operating in 7-Mode" msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml205(para) msgid "" "The NetApp NFS configuration for Data ONTAP operating in 7-Mode is an " "interface from OpenStack to Data ONTAP operating in 7-Mode storage system " "for provisioning and managing OpenStack volumes on NFS exports provided by " "the Data ONTAP operating in 7-Mode storage system which can then be accessed" " using NFS protocol." msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml211(para) msgid "" "The NFS configuration for Data ONTAP operating in 7-Mode is a direct " "interface from Cinder to the Data ONTAP operating in 7-Mode instance and as " "such does not require any additional management software to achieve the " "desired functionality. It uses NetApp ONTAPI to interact with the Data ONTAP" " operating in 7-Mode storage system." msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml218(title) msgid "" "Configuration options for the Data ONTAP operating in 7-Mode family with NFS" " protocol" msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml220(para) msgid "" "Configure the volume driver, storage family and storage protocol to the " "NetApp unified driver, Data ONTAP operating in 7-Mode, and NFS respectively " "by setting the volume_driver, " "netapp_storage_family and " "netapp_storage_protocol options in " "cinder.conf as follows:" msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml244(title) msgid "Upgrading prior NetApp drivers to the NetApp unified driver" msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml245(para) msgid "" "NetApp introduced a new unified block storage driver in Havana for " "configuring different storage families and storage protocols. This requires " "defining upgrade path for NetApp drivers which existed in releases prior to " "Havana. This section covers the upgrade configuration for NetApp drivers to " "the new unified configuration and a list of deprecated NetApp drivers." msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml253(title) msgid "Upgraded NetApp drivers" msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml254(para) msgid "" "This section describes how to update Cinder configuration from a pre-Havana " "release to the new unified driver format." msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml258(title) msgid "Driver upgrade configuration" msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml261(para) msgid "" "NetApp iSCSI direct driver for Clustered Data ONTAP in Grizzly (or earlier)" msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml266(para) #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml280(para) #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml295(para) #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml310(para) msgid "NetApp Unified Driver configuration" msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml275(para) msgid "" "NetApp NFS direct driver for Clustered Data ONTAP in Grizzly (or earlier)" msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml289(para) msgid "" "NetApp iSCSI direct driver for Data ONTAP operating in 7-Mode storage " "controller in Grizzly (or earlier)" msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml304(para) msgid "" "NetApp NFS direct driver for Data ONTAP operating in 7-Mode storage " "controller in Grizzly (or earlier)" msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml323(title) msgid "Deprecated NetApp drivers" msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml324(para) msgid "" "This section lists the NetApp drivers in previous releases that are " "deprecated in Havana." msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml328(para) msgid "NetApp iSCSI driver for clustered Data ONTAP." msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml335(para) msgid "NetApp NFS driver for clustered Data ONTAP." msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml342(para) msgid "" "NetApp iSCSI driver for Data ONTAP operating in 7-Mode storage controller." msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml349(para) msgid "" "NetApp NFS driver for Data ONTAP operating in 7-Mode storage controller." msgstr "" #: ./doc/config-reference/block-storage/drivers/netapp-volume-driver.xml357(para) msgid "" "See the OpenStack NetApp " "community for support information on deprecated NetApp drivers in the" " Havana release." msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml6(title) msgid "Huawei storage driver" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml7(para) msgid "" "Huawei driver supports the iSCSI and Fibre Channel connections and enables " "OceanStor T series unified storage, OceanStor Dorado high-performance " "storage, and OceanStor HVS high-end storage to provide block storage " "services for OpenStack." msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml20(para) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml51(para) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml77(para) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml96(para) #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml39(para) #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml33(para) msgid "Delete volume" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml23(para) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml54(para) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml80(para) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml99(para) #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml42(para) #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml36(para) msgid "Attach volume" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml29(para) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml60(para) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml105(para) #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml48(para) #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml48(para) msgid "Create snapshot" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml32(para) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml63(para) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml108(para) #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml51(para) #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml51(para) msgid "Delete snapshot" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml38(para) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml114(para) msgid "Create clone volume" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml41(para) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml66(para) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml86(para) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml117(para) #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml57(para) #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml54(para) msgid "Copy image to volume" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml44(para) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml69(para) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml89(para) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml120(para) #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml60(para) #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml57(para) msgid "Copy volume to image" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml14(para) msgid "" "OceanStor T series unified storage supports the following " "operations:OceanStor Dorado5100 supports the following " "operations:OceanStor Dorado2100 G2 supports the following " "operations:OceanStor HVS supports the following " "operations:" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml125(title) msgid "Configure Cinder nodes" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml126(para) msgid "" "In /etc/cinder, create the driver configuration file " "named cinder_huawei_conf.xml." msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml129(para) msgid "" "You must configure and to" " specify a storage system and link type. The following uses the iSCSI driver" " as an example. The driver configuration file of OceanStor T series unified " "storage is shown as follows:" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml160(para) msgid "" "The driver configuration file of OceanStor Dorado5100 is shown as follows:" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml186(para) msgid "" "The driver configuration file of OceanStor Dorado2100 G2 is shown as " "follows:" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml210(para) msgid "The driver configuration file of OceanStor HVS is shown as follows:" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml236(para) msgid "" "You do not need to configure the iSCSI target IP address for the Fibre " "Channel driver. In the prior example, delete the iSCSI configuration:" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml246(para) msgid "" "To add and " " items, you can modify the " "cinder.conf configuration file as follows:" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml252(para) msgid "You can configure multiple Huawei back-end storages as follows:" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml263(para) msgid "" "OceanStor HVS storage system supports the QoS function. You must create a " "QoS policy for the HVS storage system and create the volume type to enable " "QoS as follows:" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml272(para) msgid "" " is a QoS policy created by a user for " "the HVS storage system. is the self-defined volume" " type. Set the option to " "high, normal, or " "low." msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml280(para) msgid "" "OceanStor HVS storage system supports the SmartTier function. SmartTier has " "three tiers. You can create the volume type to enable SmartTier as follows:" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml288(para) msgid "" " and " "can only be set to high, normal, or " "low." msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml295(title) msgid "Configuration file details" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml296(para) msgid "This table describes the Huawei storage driver configuration options:" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml299(caption) msgid "Huawei storage driver configuration options" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml308(td) #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml199(td) msgid "Flag name" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml309(td) #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml200(td) #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml317(td) msgid "Type" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml310(td) #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml201(td) #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml318(td) msgid "Default" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml317(option) msgid "Product" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml320(para) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml334(td) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml345(td) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml357(para) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml369(td) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml381(para) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml396(para) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml501(td) #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml209(para) #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml224(para) #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml262(para) #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml326(para) #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml334(para) #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml379(para) #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml391(para) #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml408(para) #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml419(para) msgid "Required" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml326(para) msgid "" "Type of a storage product. Valid values are T, " "Dorado, or HVS." msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml333(option) msgid "Protocol" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml339(literal) #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml412(para) msgid "iSCSI" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml340(literal) msgid "FC" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml338(td) msgid "" "Type of a protocol. Valid values are or ." msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml344(option) msgid "ControllerIP0" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml349(td) msgid "IP address of the primary controller (not required for the HVS)" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml354(option) msgid "ControllerIP1" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml363(para) msgid "IP address of the secondary controller (not required for the HVS)" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml368(option) msgid "HVSURL" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml373(td) msgid "Access address of the Rest port (required only for the HVS)" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml378(option) msgid "UserName" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml387(para) msgid "User name of an administrator" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml392(option) msgid "UserPassword" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml402(para) msgid "Password of an administrator" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml407(option) msgid "LUNType" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml410(para) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml427(para) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml445(para) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml463(para) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml476(td) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml491(td) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml513(td) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml524(td) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml534(td) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml545(td) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml553(td) #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml217(para) #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml271(para) #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml295(para) #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml306(para) #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml330(para) #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml341(para) #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml362(para) #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml380(para) #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml390(para) #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml411(para) #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml422(para) #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml432(para) #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml448(para) #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml342(para) #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml352(para) #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml364(para) #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml430(para) #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml441(para) msgid "Optional" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml413(para) msgid "Thin" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml416(para) msgid "" "Type of a created LUN. Valid values are Thick or " "Thin." msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml423(option) msgid "StripUnitSize" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml430(para) msgid "64" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml433(para) msgid "Stripe depth of a created LUN. The value is expressed in KB." msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml435(para) msgid "This flag is not valid for a thin LUN." msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml441(option) msgid "WriteType" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml448(para) #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml466(para) msgid "1" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml451(para) msgid "" "Cache write method. The method can be write back, write through, or Required" " write back. The default value is 1, indicating write " "back." msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml460(option) msgid "MirrorSwitch" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml469(para) msgid "" "Cache mirroring policy. The default value is 1, " "indicating that a mirroring policy is used." msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml475(option) msgid "Prefetch Type" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml478(para) msgid "3" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml481(para) msgid "" "Cache prefetch strategy. The strategy can be constant prefetch, variable " "prefetch, or intelligent prefetch. Default value is 3, " "which indicates intelligent prefetch and is not required for the HVS." msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml490(option) msgid "Prefetch Value" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml493(para) #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml381(para) #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml431(para) msgid "0" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml496(para) msgid "Cache prefetch value." msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml500(option) msgid "StoragePool" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml506(para) msgid "" "Name of a storage pool that you want to use. Not required for the Dorado2100" " G2." msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml512(option) msgid "DefaultTargetIP" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml518(para) msgid "Default IP address of the iSCSI port provided for compute nodes." msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml523(option) msgid "Initiator Name" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml529(para) msgid "Name of a compute node initiator." msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml533(option) msgid "Initiator TargetIP" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml539(para) msgid "IP address of the iSCSI port provided for Compute nodes." msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml544(option) msgid "OSType" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml547(para) msgid "Linux" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml549(td) msgid "The OS type for a Compute node." msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml552(option) msgid "HostIP" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml557(td) msgid "The IPs for Compute nodes." msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml564(para) msgid "" "You can configure one iSCSI target port for each or all Compute nodes. The " "driver checks whether a target port IP address is configured for the current" " Compute node. If not, select ." msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml571(para) msgid "" "You can configure multiple storage pools in one configuration file, which " "supports the use of multiple storage pools in a storage system. (HVS allows " "configuration of only one storage pool.)" msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml578(para) msgid "" "For details about LUN configuration information, see the " "command in the command-line interface (CLI) documentation or run the " " on the storage system CLI." msgstr "" #: ./doc/config-reference/block-storage/drivers/huawei-storage-driver.xml587(para) msgid "" "After the driver is loaded, the storage system obtains any modification of " "the driver configuration file in real time and you do not need to restart " "the cinder-volume service." msgstr "" #: ./doc/config-reference/block-storage/drivers/glusterfs-driver.xml6(title) msgid "GlusterFS driver" msgstr "" #: ./doc/config-reference/block-storage/drivers/glusterfs-driver.xml7(para) msgid "" "GlusterFS is an open-source scalable distributed file system that is able to" " grow to petabytes and beyond in size. More information can be found on " "Gluster's homepage." msgstr "" #: ./doc/config-reference/block-storage/drivers/glusterfs-driver.xml12(para) msgid "" "This driver enables use of GlusterFS in a similar fashion as the NFS driver." " It supports basic volume operations, and like NFS, does not support " "snapshot/clone." msgstr "" #: ./doc/config-reference/block-storage/drivers/glusterfs-driver.xml16(para) msgid "" "You must use a Linux kernel of version 3.4 or greater (or version 2.6.32 or " "greater in RHEL/CentOS 6.3+) when working with Gluster-based volumes. See " "Bug " "1177103 for more information." msgstr "" #: ./doc/config-reference/block-storage/drivers/glusterfs-driver.xml22(para) msgid "" "To use Cinder with GlusterFS, first set the volume_driver" " in cinder.conf:" msgstr "" #: ./doc/config-reference/block-storage/drivers/glusterfs-driver.xml26(para) msgid "" "The following table contains the configuration options supported by the " "GlusterFS driver." msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml5(title) msgid "EMC SMI-S iSCSI driver" msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml6(para) msgid "" "The EMC SMI-S iSCSI driver, which is based on the iSCSI driver, can create, " "delete, attach, and detach volumes, create and delete snapshots, and so on." msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml9(para) msgid "" "The EMC SMI-S iSCSI driver runs volume operations by communicating with the " "back-end EMC storage. It uses a CIM client in Python called PyWBEM to " "perform CIM operations over HTTP." msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml13(para) msgid "" "The EMC CIM Object Manager (ECOM) is packaged with the EMC SMI-S provider. " "It is a CIM server that enables CIM clients to perform CIM operations over " "HTTP by using SMI-S in the back-end for EMC storage operations." msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml17(para) msgid "" "The EMC SMI-S Provider supports the SNIA Storage Management Initiative " "(SMI), an ANSI standard for storage management. It supports VMAX and VNX " "storage systems." msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml21(title) #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml16(title) #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml17(title) msgid "System requirements" msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml22(para) msgid "" "EMC SMI-S Provider V4.5.1 and higher is required. You can download SMI-S " "from the EMC Powerlink web " "site. See the EMC SMI-S Provider release notes for installation " "instructions." msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml27(para) msgid "EMC storage VMAX Family and VNX Series are supported." msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml32(para) msgid "VMAX and VNX arrays support these operations:" msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml54(para) msgid "Create cloned volume" msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml63(para) msgid "Only VNX supports these operations:" msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml69(para) msgid "Only thin provisioning is supported." msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml72(title) msgid "Task flow" msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml74(title) msgid "To set up the EMC SMI-S iSCSI driver" msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml76(para) msgid "" "Install the python-pywbem package for your distribution. " "See ." msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml81(para) msgid "" "Download SMI-S from PowerLink and install it. Add your VNX/VMAX arrays to " "SMI-S." msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml83(para) msgid "" "For information, see and the SMI-S release " "notes." msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml87(para) msgid "Register with VNX. See ." msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml91(para) msgid "Create a masking view on VMAX. See ." msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml96(title) msgid "Install the python-pywbem package" msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml99(para) msgid "" "Install the python-pywbem package for your distribution:" msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml103(para) msgid "On Ubuntu:" msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml107(para) msgid "On openSUSE:" msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml111(para) msgid "On Fedora:" msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml119(title) msgid "Set up SMI-S" msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml120(para) msgid "" "You can install SMI-S on a non-OpenStack host. Supported platforms include " "different flavors of Windows, Red Hat, and SUSE Linux. The host can be " "either a physical server or VM hosted by an ESX server. See the EMC SMI-S " "Provider release notes for supported platforms and installation " "instructions." msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml127(para) msgid "" "You must discover storage arrays on the SMI-S server before you can use the " "Cinder driver. Follow instructions in the SMI-S release notes." msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml131(para) msgid "" "SMI-S is usually installed at /opt/emc/ECIM/ECOM/bin on" " Linux and C:\\Program Files\\EMC\\ECIM\\ECOM\\bin on " "Windows. After you install and configure SMI-S, go to that directory and " "type ." msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml138(para) msgid "" "Use in to add an array. Use " " and examine the output after the array is added. Make sure " "that the arrays are recognized by the SMI-S server before using the EMC " "Cinder driver." msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml146(title) msgid "Register with VNX" msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml147(para) msgid "" "To export a VNX volume to a Compute node, you must register the node with " "VNX." msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml149(para) msgid "" "On the Compute node 1.1.1.1, run these commands (assume " "10.10.61.35 is the iscsi target):" msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml152(literal) #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml158(literal) msgid "10.10.61.35" msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml156(para) msgid "" "Log in to VNX from the Compute node by using the target corresponding to the" " SPA port:" msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml158(literal) msgid "iqn.1992-04.com.emc:cx.apm01234567890.a0" msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml159(para) msgid "" "Assume iqn.1993-08.org.debian:01:1a2b3c4d5f6g is the " "initiator name of the Compute node. Log in to Unisphere, go to " "VNX00000->Hosts->Initiators, refresh and wait until" " initiator iqn.1993-08.org.debian:01:1a2b3c4d5f6g with SP" " Port A-8v0 appears." msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml167(para) msgid "" "Click Register, select " "CLARiiON/VNX, and enter the myhost1 " "host name and myhost1 IP address. Click " "Register. Now the 1.1.1.1 host " "appears under HostsHost List " "as well." msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml172(para) msgid "Log out of VNX on the Compute node:" msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml174(para) msgid "" "Log in to VNX from the Compute node using the target corresponding to the " "SPB port:" msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml177(para) msgid "In Unisphere, register the initiator with the SPB port." msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml179(para) msgid "Log out:" msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml183(title) msgid "Create a masking view on VMAX" msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml184(para) msgid "" "For VMAX, you must set up the Unisphere for VMAX server. On the Unisphere " "for VMAX server, create initiator group, storage group, and port group and " "put them in a masking view. Initiator group contains the initiator names of " "the OpenStack hosts. Storage group must have at least six gatekeepers." msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml194(para) msgid "" "Make the following changes in /etc/cinder/cinder.conf." msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml196(para) msgid "" "For VMAX, add the following entries, where 10.10.61.45 is" " the IP address of the VMAX iscsi target:" msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml203(para) msgid "" "For VNX, add the following entries, where 10.10.61.35 is " "the IP address of the VNX iscsi target:" msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml210(para) msgid "" "Restart the cinder-volume " "service." msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml214(title) msgid "cinder_emc_config.xml configuration file" msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml216(para) msgid "" "Create the file /etc/cinder/cinder_emc_config.xml. You " "do not need to restart the service for this change." msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml220(para) msgid "For VMAX, add the following lines to the XML file:" msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml231(para) msgid "For VNX, add the following lines to the XML file:" msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml241(para) msgid "" "To attach VMAX volumes to an OpenStack VM, you must create a Masking View by" " using Unisphere for VMAX. The Masking View must have an Initiator Group " "that contains the initiator of the OpenStack compute node that hosts the VM." msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml246(para) msgid "" "StorageType is the thin pool where user wants to create the volume from. " "Only thin LUNs are supported by the plug-in. Thin pools can be created using" " Unisphere for VMAX and VNX." msgstr "" #: ./doc/config-reference/block-storage/drivers/emc-volume-driver.xml250(para) msgid "" "EcomServerIp and EcomServerPort are the IP address and port number of the " "ECOM server which is packaged with SMI-S. EcomUserName and EcomPassword are " "credentials for the ECOM server." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml4(title) msgid "HP 3PAR Fibre Channel and iSCSI drivers" msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml5(para) msgid "" "The HP3PARFCDriver and " "HP3PARISCSIDriver drivers, which are based on the Block" " Storage Service (Cinder) plug-in architecture, run volume operations by " "communicating with the HP 3PAR storage system over HTTP, HTTPS, and SSH " "connections. The HTTP and HTTPS communications use " "hp3parclient, which is part of the Python standard " "library." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml13(para) msgid "" "For information about how to manage HP 3PAR storage systems, see the HP 3PAR" " user documentation." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml17(para) msgid "" "To use the HP 3PAR drivers, install the following software and components on" " the HP 3PAR storage system:" msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml22(para) msgid "HP 3PAR Operating System software version 3.1.2 (MU2) or higher" msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml26(para) msgid "HP 3PAR Web Services API Server must be enabled and running" msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml30(para) msgid "One Common Provisioning Group (CPG)" msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml33(para) msgid "" "Additionally, you must install the hp3parclient version " "2.0 or newer from the Python standard library on the system with the enabled" " Block Storage Service volume drivers." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml45(para) msgid "Create volumes." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml48(para) msgid "Delete volumes." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml51(para) msgid "Extend volumes." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml54(para) msgid "Attach volumes." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml57(para) msgid "Detach volumes." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml60(para) msgid "Create snapshots." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml63(para) msgid "Delete snapshots." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml66(para) msgid "Create volumes from snapshots." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml69(para) msgid "Create cloned volumes." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml72(para) msgid "Copy images to volumes." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml75(para) msgid "Copy volumes to images." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml78(para) msgid "" "Volume type support for both HP 3PAR drivers includes the ability to set the" " following capabilities in the OpenStack Cinder API " "cinder.api.contrib.types_extra_specs volume type extra " "specs extension module:" msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml85(literal) msgid "hp3par:cpg" msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml88(literal) msgid "hp3par:snap_cpg" msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml91(literal) msgid "hp3par:provisioning" msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml94(literal) msgid "hp3par:persona" msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml97(literal) msgid "hp3par:vvs" msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml100(literal) msgid "qos:maxBWS" msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml103(literal) msgid "qos:maxIOPS" msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml106(para) msgid "" "To work with the default filter scheduler, the key values are case sensitive" " and scoped with hp3par: or qos:. For " "information about how to set the key-value pairs and associate them with a " "volume type, run the following command: " msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml116(para) msgid "" "Volumes that are cloned only support extra specs keys cpg, snap_cpg, " "provisioning and vvs. The others are ignored. In addition the comments " "section of the cloned volume in the HP 3PAR StoreServ storage array is not " "populated." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml122(para) msgid "" "The following keys require that the HP 3PAR StoreServ storage array has a " "Priority Optimization license installed." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml127(para) msgid "" "hp3par:vvs - The virtual volume set name that has been " "predefined by the Administrator with Quality of Service (QoS) rules " "associated to it. If you specify hp3par:vvs, the " "qos:maxIOPS and qos:maxBWS settings " "are ignored." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml137(para) msgid "" "qos:maxBWS - The QoS I/O issue count rate limit in MBs. " "If not set, the I/O issue bandwidth rate has no limit." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml142(para) msgid "" "qos:maxIOPS - The QoS I/O issue count rate limit. If not " "set, the I/O issue count rate has no limit." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml147(para) msgid "" "If volume types are not used or a particular key is not set for a volume " "type, the following defaults are used." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml152(para) msgid "" "hp3par:cpg - Defaults to the " "hp3par_cpg setting in the " "cinder.conf file." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml157(para) msgid "" "hp3par:snap_cpg - Defaults to the " "hp3par_snap setting in the " "cinder.conf file. If hp3par_snap is " "not set, it defaults to the hp3par_cpg setting." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml165(para) msgid "" "hp3par:provisioning - Defaults to thin provisioning, the " "valid values are thin and full." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml171(para) msgid "" "hp3par:persona - Defaults to the 1 – " "Generic persona. The valid values are, 1 – " "Generic, 2 - Generic-ALUA, 6 - " "Generic-legacy, 7 - HPUX-legacy, 8 - " "AIX-legacy, 9 – EGENERA, 10 - ONTAP-" "legacy, 11 – VMware, and 12 - " "OpenVMS." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml186(title) msgid "Enable the HP 3PAR Fibre Channel and iSCSI drivers" msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml188(para) msgid "" "The HP3PARFCDriver and " "HP3PARISCSIDriver are installed with the OpenStack " "software." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml193(para) msgid "" "Install the hp3parclient Python package on the " "OpenStack Block Storage system. " msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml200(para) msgid "" "Verify that the HP 3PAR Web Services API server is enabled and running on " "the HP 3PAR storage system." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml205(para) msgid "" "Log onto the HP 3PAR storage system with administrator " "access." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml210(para) msgid "" "View the current state of the Web Services API Server. " "" msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml217(para) msgid "If the Web Services API Server is disabled, start it." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml227(para) msgid "" "To stop the Web Services API Server, use the stopwsapi command. For other " "options run the command." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml224(para) msgid "" "If the HTTP or HTTPS state is disabled, enable one of them. " "or " msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml234(para) msgid "" "If you are not using an existing CPG, create a CPG on the HP 3PAR storage " "system to be used as the default location for creating volumes." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml239(para) msgid "" "Make the following changes in the " "/etc/cinder/cinder.conf file." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml243(emphasis) msgid "## REQUIRED SETTINGS" msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml277(emphasis) msgid "## OPTIONAL SETTINGS" msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml291(para) msgid "" "You can enable only one driver on each cinder instance unless you enable " "multiple back-end support. See the Cinder multiple back-end support " "instructions to enable this feature." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml298(para) msgid "" "You can configure one or more iSCSI addresses by using the " " option. When you configure multiple " "addresses, the driver selects the iSCSI port with the fewest active volumes " "at attach time. The IP address might include an IP port by using a colon " "(:) to separate the address from port. If you do not " "define an IP port, the default port 3260 is used. Separate IP addresses with" " a comma (,). The " "/ options might " "be used as an alternative to for single " "port iSCSI configuration." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml317(para) msgid "" "Save the changes to the cinder.conf file and restart " "the cinder-volume service." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-3par-driver.xml323(para) msgid "" "The HP 3PAR Fibre Channel and iSCSI drivers are now enabled on your " "OpenStack system. If you experience problems, review the Block Storage " "Service log files for errors." msgstr "" #: ./doc/config-reference/block-storage/drivers/dell-equallogic-driver.xml6(title) msgid "Dell EqualLogic volume driver" msgstr "" #: ./doc/config-reference/block-storage/drivers/dell-equallogic-driver.xml7(para) msgid "" "The Dell EqualLogic volume driver interacts with configured EqualLogic " "arrays and supports various operations, such as volume creation and " "deletion, volume attachment and detachment, snapshot creation and deletion, " "and clone creation." msgstr "" #: ./doc/config-reference/block-storage/drivers/dell-equallogic-driver.xml12(para) msgid "" "To configure and use a Dell EqualLogic array with Block Storage, modify your" " cinder.conf as follows." msgstr "" #: ./doc/config-reference/block-storage/drivers/dell-equallogic-driver.xml15(para) msgid "" "Set the option to the Dell EqualLogic volume " "driver:" msgstr "" #: ./doc/config-reference/block-storage/drivers/dell-equallogic-driver.xml18(para) msgid "" "Set the option to the IP address to reach the " "EqualLogic Group through SSH:" msgstr "" #: ./doc/config-reference/block-storage/drivers/dell-equallogic-driver.xml21(para) msgid "" "Set the option to the user name to login to the " "Group manager:" msgstr "" #: ./doc/config-reference/block-storage/drivers/dell-equallogic-driver.xml24(para) msgid "" "Set the option to the password to login the " "Group manager with:" msgstr "" #: ./doc/config-reference/block-storage/drivers/dell-equallogic-driver.xml27(para) msgid "" "Optionally set the option to false to " "disable creation of thin-provisioned volumes:" msgstr "" #: ./doc/config-reference/block-storage/drivers/dell-equallogic-driver.xml31(para) msgid "" "The following table describes additional options that the driver supports:" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml5(title) msgid "IBM Storwize family and SVC volume driver" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml6(para) msgid "" "The volume management driver for Storwize family and SAN Volume Controller " "(SVC) provides OpenStack Compute instances with access to IBM Storwize " "family or SVC storage systems." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml11(title) msgid "Configure the Storwize family and SVC system" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml13(title) msgid "Network configuration" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml14(para) msgid "" "The Storwize family or SVC system must be configured for iSCSI, Fibre " "Channel, or both." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml16(para) msgid "" "If using iSCSI, each Storwize family or SVC node should have at least one " "iSCSI IP address. The IBM Storwize/SVC driver uses an iSCSI IP address " "associated with the volume's preferred node (if available) to attach the " "volume to the instance, otherwise it uses the first available iSCSI IP " "address of the system. The driver obtains the iSCSI IP address directly from" " the storage system; you do not need to provide these iSCSI IP addresses " "directly to the driver." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml27(para) msgid "" "If using iSCSI, ensure that the compute nodes have iSCSI network access to " "the Storwize family or SVC system." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml32(para) msgid "" "OpenStack Nova's Grizzly version supports iSCSI multipath. Once this is " "configured on the Nova host (outside the scope of this documentation), " "multipath is enabled." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml37(para) msgid "" "If using Fibre Channel (FC), each Storwize family or SVC node should have at" " least one WWPN port configured. If the " "storwize_svc_multipath_enabled flag is set to True in the" " Cinder configuration file, the driver uses all available WWPNs to attach " "the volume to the instance (details about the configuration flags appear in " "the next section). If the" " flag is not set, the driver uses the WWPN associated with the volume's " "preferred node (if available), otherwise it uses the first available WWPN of" " the system. The driver obtains the WWPNs directly from the storage system; " "you do not need to provide these WWPNs directly to the driver." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml53(para) msgid "" "If using FC, ensure that the compute nodes have FC connectivity to the " "Storwize family or SVC system." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml59(title) msgid "iSCSI CHAP authentication" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml67(para) msgid "" "CHAP secrets are added to existing hosts as well as newly-created ones. If " "the CHAP option is enabled, hosts will not be able to access the storage " "without the generated secrets." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml74(para) msgid "" "Not all OpenStack Compute drivers support CHAP authentication. Please check " "compatibility before using." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml79(para) msgid "" "CHAP secrets are passed from OpenStack Block Storage to Compute in clear " "text. This communication should be secured to ensure that CHAP secrets are " "not discovered." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml60(para) msgid "" "If using iSCSI for data access and the " "storwize_svc_iscsi_chap_enabled is set to " "True, the driver will associate randomly-generated CHAP " "secrets with all hosts on the Storwize family system. OpenStack compute " "nodes use these secrets when creating iSCSI connections. " "" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml86(title) msgid "Configure storage pools" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml87(para) msgid "" "Each instance of the IBM Storwize/SVC driver allocates all volumes in a " "single pool. The pool should be created in advance and be provided to the " "driver using the storwize_svc_volpool_name configuration " "flag. Details about the configuration flags and how to provide the flags to " "the driver appear in the next " "section." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml98(title) msgid "Configure user authentication for the driver" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml100(para) msgid "" "The driver requires access to the Storwize family or SVC system management " "interface. The driver communicates with the management using SSH. The driver" " should be provided with the Storwize family or SVC management IP using the " "san_ip flag, and the management port should be provided " "by the san_ssh_port flag. By default, the port value is " "configured to be port 22 (SSH)." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml110(para) msgid "" "Make sure the compute node running the nova-volume management driver has SSH" " network access to the storage system." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml114(para) msgid "" "To allow the driver to communicate with the Storwize family or SVC system, " "you must provide the driver with a user on the storage system. The driver " "has two authentication methods: password-based authentication and SSH key " "pair authentication. The user should have an Administrator role. It is " "suggested to create a new user for the management driver. Please consult " "with your storage and security administrator regarding the preferred " "authentication method and how passwords or SSH keys should be stored in a " "secure manner." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml125(para) msgid "" "When creating a new user on the Storwize or SVC system, make sure the user " "belongs to the Administrator group or to another group that has an " "Administrator role." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml130(para) msgid "" "If using password authentication, assign a password to the user on the " "Storwize or SVC system. The driver configuration flags for the user and " "password are san_login and " "san_password, respectively." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml136(para) msgid "" "If you are using the SSH key pair authentication, create SSH private and " "public keys using the instructions below or by any other method. Associate " "the public key with the user by uploading the public key: select the " "\"choose file\" option in the Storwize family or SVC management GUI under " "\"SSH public key\". Alternatively, you may associate the SSH public key " "using the command line interface; details can be found in the Storwize and " "SVC documentation. The private key should be provided to the driver using " "the san_private_key configuration flag." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml150(title) msgid "Create a SSH key pair with OpenSSH" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml151(para) msgid "You can create an SSH key pair using OpenSSH, by running:" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml154(para) msgid "" "The command prompts for a file to save the key pair. For example, if you " "select 'key' as the filename, two files are created: key " "and key.pub. The key file holds the " "private SSH key and key.pub holds the public SSH key." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml161(para) msgid "The command also prompts for a pass phrase, which should be empty." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml163(para) msgid "" "The private key file should be provided to the driver using the " "san_private_key configuration flag. The public key should" " be uploaded to the Storwize family or SVC system using the storage " "management GUI or command line interface." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml169(para) msgid "Ensure that Cinder has read permissions on the private key file." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml175(title) msgid "Configure the Storwize family and SVC driver" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml177(title) msgid "Enable the Storwize family and SVC driver" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml178(para) msgid "" "Set the volume driver to the Storwize family and SVC driver by setting the " "volume_driver option in cinder.conf " "as follows:" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml185(title) msgid "Storwize family and SVC driver options in cinder.conf" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml187(para) msgid "" "The following options specify default values for all volumes. Some can be " "over-ridden using volume types, which are described below." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml191(caption) msgid "List of configuration flags for Storwize storage and SVC driver" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml207(literal) msgid "san_ip" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml211(para) msgid "Management IP or host name" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml215(literal) msgid "san_ssh_port" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml218(para) msgid "22" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml219(para) msgid "Management port" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml222(literal) msgid "san_login" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml226(para) msgid "Management login username" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml230(literal) msgid "san_password" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml234(para) msgid "" "The authentication requires either a password " "(san_password) or SSH private key " "(san_private_key). One must be specified. If both are " "specified, the driver uses only the SSH private key." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml232(para) msgid "Required " msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml244(para) msgid "Management login password" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml248(literal) msgid "san_private_key" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml250(para) msgid "Required " msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml254(para) msgid "Management login SSH private key" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml260(literal) msgid "storwize_svc_volpool_name" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml264(para) msgid "Default pool name for volumes" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml269(literal) msgid "storwize_svc_vol_rsize" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml272(para) msgid "2" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml276(para) msgid "" "The driver creates thin-provisioned volumes by default. The " "storwize_svc_vol_rsize flag defines the initial physical " "allocation percentage for thin-provisioned volumes, or if set to " "-1, the driver creates full allocated volumes. More " "details about the available options are available in the Storwize family and" " SVC documentation." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml273(para) msgid "Initial physical allocation (percentage) " msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml293(literal) msgid "storwize_svc_vol_warning" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml296(para) msgid "0 (disabled)" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml297(para) msgid "" "Space allocation warning threshold (percentage) " msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml304(literal) msgid "storwize_svc_vol_autoexpand" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml307(para) #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml363(para) #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml423(para) #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml449(para) msgid "True" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml310(para) msgid "" "Defines whether thin-provisioned volumes can be auto expanded by the storage" " system, a value of True means that auto expansion is " "enabled, a value of False disables auto expansion. " "Details about this option can be found in the –autoexpand" " flag of the Storwize family and SVC command line interface " "mkvdisk command." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml308(para) msgid "Enable or disable volume auto expand " msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml328(literal) msgid "storwize_svc_vol_grainsize" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml331(para) msgid "256" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml332(para) msgid "Volume grain size in KB" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml338(literal) msgid "storwize_svc_vol_compression" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml342(para) #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml433(para) msgid "False" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml346(para) msgid "" "Defines whether Real-time Compression is used for the volumes created with " "OpenStack. Details on Real-time Compression can be found in the Storwize " "family and SVC documentation. The Storwize or SVC system must have " "compression enabled for this feature to work." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml343(para) msgid "Enable or disable Real-time Compression " msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml360(literal) msgid "storwize_svc_vol_easytier" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml366(para) msgid "" "Defines whether Easy Tier is used for the volumes created with OpenStack. " "Details on EasyTier can be found in the Storwize family and SVC " "documentation. The Storwize or SVC system must have Easy Tier enabled for " "this feature to work." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml364(para) msgid "Enable or disable Easy Tier " msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml378(literal) msgid "storwize_svc_vol_iogrp" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml382(para) msgid "The I/O group in which to allocate vdisks" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml387(literal) msgid "storwize_svc_flashcopy_timeout" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml391(para) msgid "120" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml394(para) msgid "" "The driver wait timeout threshold when creating an OpenStack snapshot. This " "is actually the maximum amount of time that the driver waits for the " "Storwize family or SVC system to prepare a new FlashCopy mapping. The driver" " accepts a maximum wait time of 600 seconds (10 minutes)." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml392(para) msgid "FlashCopy timeout threshold (seconds)" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml408(literal) msgid "storwize_svc_connection_protocol" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml413(para) msgid "Connection protocol to use (currently supports 'iSCSI' or 'FC')" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml419(literal) msgid "storwize_svc_iscsi_chap_enabled" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml424(para) msgid "Configure CHAP authentication for iSCSI connections" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml429(literal) msgid "storwize_svc_multipath_enabled" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml436(para) msgid "" "Multipath for iSCSI connections requires no storage-side configuration and " "is enabled if the compute host has multipath configured." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml434(para) msgid "Enable multipath for FC connections " msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml445(literal) msgid "storwize_svc_multihost_enabled" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml453(para) msgid "" "This option allows the driver to map a vdisk to more than one host at a " "time. This scenario occurs during migration of a virtual machine with an " "attached volume; the volume is simultaneously mapped to both the source and " "destination compute hosts. If your deployment does not require attaching " "vdisks to multiple hosts, setting this flag to False will provide added " "safety." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml450(para) msgid "Enable mapping vdisks to multiple hosts " msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml474(title) msgid "Placement with volume types" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml475(para) msgid "" "The IBM Storwize/SVC driver exposes capabilities that can be added to the " "extra specs of volume types, and used by the filter " "scheduler to determine placement of new volumes. Make sure to prefix these " "keys with capabilities: to indicate that the scheduler " "should use them. The following extra specs are supported:" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml486(para) msgid "" "capabilities:volume_back-end_name - Specify a specific back-end where the " "volume should be created. The back-end name is a concatenation of the name " "of the IBM Storwize/SVC storage system as shown in " "lssystem, an underscore, and the name of the pool (mdisk " "group). For example: " msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml497(para) msgid "" "capabilities:compression_support - Specify a back-end according to " "compression support. A value of True should be used to " "request a back-end that supports compression, and a value of " "False will request a back-end that does not support " "compression. If you do not have constraints on compression support, do not " "set this key. Note that specifying True does not enable " "compression; it only requests that the volume be placed on a back-end that " "supports compression. Example syntax: " msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml513(para) msgid "" "capabilities:easytier_support - Similar semantics as the " "compression_support key, but for specifying according to " "support of the Easy Tier feature. Example syntax: " msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml521(para) msgid "" "capabilities:storage_protocol - Specifies the connection protocol used to " "attach volumes of this type to instances. Legal values are " "iSCSI and FC. This extra " "specs value is used for both placement and setting the protocol " "used for this volume. In the example syntax, note <in> is used as " "opposed to <is> used in the previous examples. " msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml536(title) msgid "Configure per-volume creation options" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml537(para) msgid "" "Volume types can also be used to pass options to the IBM Storwize/SVC " "driver, which over-ride the default values set in the configuration file. " "Contrary to the previous examples where the \"capabilities\" scope was used " "to pass parameters to the Cinder scheduler, options can be passed to the IBM" " Storwize/SVC driver with the \"drivers\" scope." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml547(para) msgid "rsize" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml550(para) msgid "warning" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml553(para) msgid "autoexpand" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml556(para) msgid "grainsize" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml559(para) msgid "compression" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml562(para) msgid "easytier" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml565(para) msgid "multipath" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml568(para) msgid "iogrp" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml544(para) msgid "" "The following extra specs keys are supported by the IBM " "Storwize/SVC driver: These keys have the same semantics as " "their counterparts in the configuration file. They are set similarly; for " "example, rsize=2 or compression=False." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml576(title) msgid "Example: Volume types" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml577(para) msgid "" "In the following example, we create a volume type to specify a controller " "that supports iSCSI and compression, to use iSCSI when attaching the volume," " and to enable compression:" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml583(para) msgid "We can then create a 50GB volume using this type:" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml587(para) msgid "Volume types can be used, for example, to provide users with different" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml591(para) msgid "" "performance levels (such as, allocating entirely on an HDD tier, using Easy " "Tier for an HDD-SDD mix, or allocating entirely on an SSD tier)" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml597(para) msgid "" "resiliency levels (such as, allocating volumes in pools with different RAID " "levels)" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml602(para) msgid "features (such as, enabling/disabling Real-time Compression)" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml609(title) msgid "Operational notes for the Storwize family and SVC driver" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml612(title) msgid "Migrate volumes" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml613(para) msgid "" "In the context of OpenStack Block Storage's volume migration feature, the " "IBM Storwize/SVC driver enables the storage's virtualization technology. " "When migrating a volume from one pool to another, the volume will appear in " "the destination pool almost immediately, while the storage moves the data in" " the background." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml621(para) msgid "" "To enable this feature, both pools involved in a given volume migration must" " have the same values for extent_size. If the pools have " "different values for extent_size, the data will still be " "moved directly between the pools (not host-side copy), but the operation " "will be synchronous." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml632(title) msgid "Extend volumes" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml633(para) msgid "" "The IBM Storwize/SVC driver allows for extending a volume's size, but only " "for volumes without snapshots." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml638(title) #: ./doc/config-reference/block-storage/drivers/ibm-gpfs-volume-driver.xml173(title) msgid "Snapshots and clones" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-storwize-svc-driver.xml639(para) msgid "" "Snapshots are implemented using FlashCopy with no background copy (space-" "efficient). Volume clones (volumes created from existing volumes) are " "implemented with FlashCopy, but with background copy enabled. This means " "that volume clones are independent, full copies. While this background copy " "is taking place, attempting to delete or extend the source volume will " "result in that operation waiting for the copy to complete." msgstr "" #. When image changes, this message will be marked fuzzy or untranslated for #. you. #. It doesn't matter what you translate it to: it's not used at all. #: ./doc/config-reference/block-storage/drivers/xenapi-nfs.xml62(None) msgid "" "@@image: '../../../common/figures/xenapinfs/local_config.png'; " "md5=16a3864b0ec636518335246360438fd1" msgstr "" #. When image changes, this message will be marked fuzzy or untranslated for #. you. #. It doesn't matter what you translate it to: it's not used at all. #: ./doc/config-reference/block-storage/drivers/xenapi-nfs.xml78(None) msgid "" "@@image: '../../../common/figures/xenapinfs/remote_config.png'; " "md5=eab22f6aa5413c2043936872ea44e459" msgstr "" #: ./doc/config-reference/block-storage/drivers/xenapi-nfs.xml4(title) msgid "XenAPINFS" msgstr "" #: ./doc/config-reference/block-storage/drivers/xenapi-nfs.xml5(para) msgid "" "XenAPINFS is a Block Storage (Cinder) driver that uses an NFS share through " "the XenAPI Storage Manager to store virtual disk images and expose those " "virtual disks as volumes." msgstr "" #: ./doc/config-reference/block-storage/drivers/xenapi-nfs.xml8(para) msgid "" "This driver does not access the NFS share directly. It accesses the share " "only through XenAPI Storage Manager. Consider this driver as a reference " "implementation for use of the XenAPI Storage Manager in OpenStack (present " "in XenServer and XCP)." msgstr "" #: ./doc/config-reference/block-storage/drivers/xenapi-nfs.xml17(para) msgid "" "A XenServer/XCP installation that acts as Storage Controller. This " "hypervisor is known as the storage controller." msgstr "" #: ./doc/config-reference/block-storage/drivers/xenapi-nfs.xml22(para) msgid "Use XenServer/XCP as your hypervisor for Compute nodes." msgstr "" #: ./doc/config-reference/block-storage/drivers/xenapi-nfs.xml26(para) msgid "" "An NFS share that is configured for XenServer/XCP. For specific requirements" " and export options, see the administration guide for your specific " "XenServer version. The NFS share must be accessible by all XenServers " "components within your cloud." msgstr "" #: ./doc/config-reference/block-storage/drivers/xenapi-nfs.xml34(para) msgid "" "To create volumes from XenServer type images (vhd tgz files), XenServer Nova" " plug-ins are also required on the storage controller." msgstr "" #: ./doc/config-reference/block-storage/drivers/xenapi-nfs.xml40(para) msgid "" "You can use a XenServer as a storage controller and Compute node at the same" " time. This minimal configuration consists of a XenServer/XCP box and an NFS" " share." msgstr "" #: ./doc/config-reference/block-storage/drivers/xenapi-nfs.xml47(title) msgid "Configuration patterns" msgstr "" #: ./doc/config-reference/block-storage/drivers/xenapi-nfs.xml50(para) msgid "" "Local configuration (Recommended): The driver runs in a virtual machine on " "top of the storage controller. With this configuration, you can create " "volumes from qemu-img-supported formats." msgstr "" #: ./doc/config-reference/block-storage/drivers/xenapi-nfs.xml57(title) msgid "Local configuration" msgstr "" #: ./doc/config-reference/block-storage/drivers/xenapi-nfs.xml68(para) msgid "" "Remote configuration: The driver is not a guest VM of the storage " "controller. With this configuration, you can only use XenServer vhd-type " "images to create volumes." msgstr "" #: ./doc/config-reference/block-storage/drivers/xenapi-nfs.xml73(title) msgid "Remote configuration" msgstr "" #: ./doc/config-reference/block-storage/drivers/xenapi-nfs.xml87(para) msgid "Assuming the following setup:" msgstr "" #: ./doc/config-reference/block-storage/drivers/xenapi-nfs.xml90(para) msgid "XenServer box at 10.2.2.1" msgstr "" #: ./doc/config-reference/block-storage/drivers/xenapi-nfs.xml94(para) msgid "XenServer password is r00tme" msgstr "" #: ./doc/config-reference/block-storage/drivers/xenapi-nfs.xml98(para) msgid "NFS server is nfs.example.com" msgstr "" #: ./doc/config-reference/block-storage/drivers/xenapi-nfs.xml102(para) msgid "NFS export is at /volumes" msgstr "" #: ./doc/config-reference/block-storage/drivers/xenapi-nfs.xml106(para) msgid "" "To use XenAPINFS as your cinder driver, set these configuration options in " "the cinder.conf file:" msgstr "" #: ./doc/config-reference/block-storage/drivers/xenapi-nfs.xml115(para) msgid "" "The following table shows the configuration options that the XenAPINFS " "driver supports:" msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml11(title) msgid "HDS iSCSI volume driver" msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml12(para) msgid "" "This Cinder volume driver provides iSCSI support for HUS (Hitachi Unified Storage) arrays such " "as, HUS-110, HUS-130, and HUS-150." msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml18(para) msgid "" "Use the HDS command to communicate with an HUS array. You " "can download this utility package from the HDS support site (https://HDSSupport.hds.com." msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml23(para) msgid "Platform: Ubuntu 12.04LTS or newer." msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml26(title) msgid "Supported Cinder operations" msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml27(para) msgid "These operations are supported:" msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml42(para) msgid "Clone volume" msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml45(para) msgid "Extend volume" msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml63(para) msgid "get_volume_stats" msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml66(para) msgid "" "Thin provisioning, also known as Hitachi Dynamic Pool (HDP), is supported " "for volume or snapshot creation. Cinder volumes and snapshots do not have to" " reside in the same pool." msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml78(para) msgid "Do not confuse differentiated services with the Cinder volume service." msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml73(para) msgid "" "The HDS driver supports the concept of differentiated services, where volume" " type can be associated with the fine-tuned performance characteristics of " "HDP—the the dynamic pool where volumes shall be created. For" " instance, an HDP can consist of fast SSDs to provide speed. HDP can provide" " a certain reliability based on things like its RAID level characteristics. " "HDS driver maps volume type to the option in " "its configuration file." msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml86(para) msgid "" "Configuration is read from an XML-format file. Examples are shown for single" " and multi back-end cases." msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml91(para) msgid "" "Configuration is read from an XML file. This example shows the configuration" " for single back-end and for multi-back-end cases." msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml100(para) msgid "" "It is okay to manage multiple HUS arrays by using multiple Cinder instances " "(or servers)." msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml96(para) msgid "" "It is not recommended to manage a HUS array simultaneously from multiple " "Cinder instances or servers. " msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml109(title) msgid "Single back-end" msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml110(para) msgid "" "In a single back-end deployment, only one Cinder instance runs on the Cinder" " server and controls one HUS array: this setup requires these configuration " "files:" msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml123(para) msgid "The configuration file location is not fixed." msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml116(para) msgid "" "Set the option in the " "/etc/cinder/cinder.conf file to use the HDS volume " "driver. This option points to a configuration file." msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml130(para) msgid "" "Configure at the location specified " "previously. For example, " "/opt/hds/hus/cinder_hds_conf.xml:" msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml160(title) msgid "Multi back-end" msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml161(para) msgid "" "In a multi back-end deployment, more than one Cinder instance runs on the " "same server. In this example, two HUS arrays are used, possibly providing " "different storage performance:" msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml167(para) msgid "" "Configure /etc/cinder/cinder.conf: the " "hus1 configuration blocks are " "created. Set the option to point to " "an unique configuration file for each block. Set the " " option for each back-end to " "cinder.volume.drivers.hds.hds.HUSDriver" msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml189(para) msgid "Configure /opt/hds/hus/cinder_hus1_conf.xml:" msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml214(para) msgid "" "Configure the /opt/hds/hus/cinder_hus2_conf.xml file:" msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml242(title) msgid "Type extra specs: and volume type" msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml244(para) msgid "" "If you use volume types, you must configure them in the configuration file " "and set the option to the appropriate " "back-end. In the previous multi back-end example, the " "platinum volume type is served by hus-2, and the " "regular volume type is served by hus-1." msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml255(title) msgid "Non differentiated deployment of HUS arrays" msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml256(para) msgid "" "You can deploy multiple Cinder instances that each control a separate HUS " "array. Each instance has no volume type associated with it. The Cinder " "filtering algorithm selects the HUS array with the largest available free " "space. In each configuration file, you must define the " "default in the service " "labels." msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml267(title) msgid "HDS iSCSI volume driver configuration options" msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml274(para) msgid "There is no relative precedence or weight among these four labels." msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml268(para) msgid "" "These details apply to the XML format configuration file that is read by HDS" " volume driver. These differentiated service labels are predefined: " "svc_0, svc_1, " "svc_2, and svc_3. Each" " respective service label associates with these parameters and tags:" msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml280(para) msgid "" ": A create_volume call with a certain volume " "type shall be matched up with this tag. default is " "special in that any service associated with this type is used to create " "volume when no other labels match. Other labels are case sensitive and " "should exactly match. If no configured volume_types match the incoming " "requested type, an error occurs in volume creation." msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml291(para) msgid ", the pool ID associated with the service." msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml295(para) msgid "An iSCSI port dedicated to the service." msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml304(para) msgid "" "get_volume_stats() always provides the available capacity based on the " "combined sum of all the HDPs that are used in these services labels." msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml298(para) msgid "" "Typically a Cinder volume instance has only one such service label. For " "example, any svc_0, svc_1, " "svc_2, or svc_3 can be associated with" " it. But any mix of these service labels can be used in the same instance " "." msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml316(td) msgid "Option" msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml324(option) msgid "mgmt_ip0" msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml328(para) msgid "Management Port 0 IP address" msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml332(option) msgid "mgmt_ip1" msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml336(para) msgid "Management Port 1 IP address" msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml345(para) msgid "Username is required only if secure mode is used" msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml355(para) msgid "Password is required only if secure mode is used" msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml361(option) msgid "svc_0, svc_1, svc_2, svc_3" msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml365(para) msgid "(at least one label has to be defined)" msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml368(para) msgid "" "Service labels: these four predefined names help four different sets of " "configuration options -- each can specify iSCSI port address, HDP and an " "unique volume type." msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml377(option) msgid "snapshot" msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml382(para) msgid "" "A service label which helps specify configuration for snapshots, such as, " "HDP." msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml389(option) msgid "volume_type" msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml394(para) msgid "" " tag is used to match volume type. " "Default meets any type of , " "or if it is not specified. Any other volume_type is selected if exactly " "matched during create_volume." msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml406(option) msgid "iscsi_ip" msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml411(para) msgid "iSCSI port IP address where volume attaches for this volume type." msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml417(option) msgid "hdp" msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml422(para) msgid "HDP, the pool number where volume, or snapshot should be created." msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml428(option) msgid "lun_start" msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml433(para) msgid "LUN allocation starts at this number." msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml439(option) msgid "lun_end" msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml442(para) msgid "4096" msgstr "" #: ./doc/config-reference/block-storage/drivers/hds-volume-driver.xml444(para) msgid "LUN allocation is up to, but not including, this number." msgstr "" #. When image changes, this message will be marked fuzzy or untranslated for #. you. #. It doesn't matter what you translate it to: it's not used at all. #: ./doc/config-reference/block-storage/drivers/ceph-rbd-volume-driver.xml26(None) msgid "" "@@image: '../../../common/figures/ceph/ceph-architecture.png'; " "md5=f7e854c9dbfb64534c47c3583e774c81" msgstr "" #: ./doc/config-reference/block-storage/drivers/ceph-rbd-volume-driver.xml4(title) msgid "Ceph RADOS Block Device (RBD)" msgstr "" #: ./doc/config-reference/block-storage/drivers/ceph-rbd-volume-driver.xml5(para) msgid "" "By Sebastien Han from " msgstr "" #: ./doc/config-reference/block-storage/drivers/ceph-rbd-volume-driver.xml8(para) msgid "" "If you use KVM or QEMU as your hypervisor, you can configure the Compute " "service to use " "Ceph RADOS block devices (RBD) for volumes." msgstr "" #: ./doc/config-reference/block-storage/drivers/ceph-rbd-volume-driver.xml21(title) msgid "Ceph architecture" msgstr "" #: ./doc/config-reference/block-storage/drivers/ceph-rbd-volume-driver.xml12(para) msgid "" "Ceph is a massively scalable, open source, distributed storage system. It is" " comprised of an object store, block store, and a POSIX-compliant " "distributed file system. The platform can auto-scale to the exabyte level " "and beyond. It runs on commodity hardware, is self-healing and self-" "managing, and has no single point of failure. Ceph is in the Linux kernel " "and is integrated with the OpenStack cloud operating system. Due to its open" " source nature, you can install and use this portable storage platform in " "public or private clouds. " msgstr "" #: ./doc/config-reference/block-storage/drivers/ceph-rbd-volume-driver.xml32(title) msgid "RADOS?" msgstr "" #: ./doc/config-reference/block-storage/drivers/ceph-rbd-volume-driver.xml33(para) msgid "You can easily get confused by the naming: Ceph? RADOS?" msgstr "" #: ./doc/config-reference/block-storage/drivers/ceph-rbd-volume-driver.xml35(para) msgid "" "RADOS: Reliable Autonomic Distributed Object Store is " "an object store. RADOS distributes objects across the storage cluster and " "replicates objects for fault tolerance. RADOS contains the following major " "components:" msgstr "" #: ./doc/config-reference/block-storage/drivers/ceph-rbd-volume-driver.xml42(para) msgid "" "Object Storage Device (ODS). The storage daemon - RADOS" " service, the location of your data. You must run this daemon on each server" " in your cluster. For each OSD, you can have an associated hard drive disks." " For performance purposes, pool your hard drive disk with raid arrays, " "logical volume management (LVM) or B-tree file system (Btrfs) pooling. By " "default, the following pools are created: data, metadata, and RBD." msgstr "" #: ./doc/config-reference/block-storage/drivers/ceph-rbd-volume-driver.xml54(para) msgid "" "Meta-Data Server (MDS). Stores metadata. MDSs build a " "POSIX file system on top of objects for Ceph clients. However, if you do not" " use the Ceph file system, you do not need a metadata server." msgstr "" #: ./doc/config-reference/block-storage/drivers/ceph-rbd-volume-driver.xml61(para) msgid "" "Monitor (MON). This lightweight daemon handles all " "communications with external applications and clients. It also provides a " "consensus for distributed decision making in a Ceph/RADOS cluster. For " "instance, when you mount a Ceph shared on a client, you point to the address" " of a MON server. It checks the state and the consistency of the data. In an" " ideal setup, you must run at least three ceph-mon daemons on " "separate servers." msgstr "" #: ./doc/config-reference/block-storage/drivers/ceph-rbd-volume-driver.xml74(para) msgid "" "Ceph developers recommend that you use Btrfs as a file system for storage. " "XFS might be a better alternative for production environments. Neither Ceph " "nor Btrfs is ready for production and it could be risky to use them in " "combination. XFS is an excellent alternative to Btrfs. The ext4 file system " "is also compatible but does not exploit the power of Ceph." msgstr "" #: ./doc/config-reference/block-storage/drivers/ceph-rbd-volume-driver.xml82(para) msgid "" "Currently, configure Ceph to use the XFS file system. Use Btrfs when it is " "stable enough for production." msgstr "" #: ./doc/config-reference/block-storage/drivers/ceph-rbd-volume-driver.xml86(para) msgid "" "See ceph.com/docs/master/rec/file" " system/ for more information about usable file systems." msgstr "" #: ./doc/config-reference/block-storage/drivers/ceph-rbd-volume-driver.xml92(title) msgid "Ways to store, use, and expose data" msgstr "" #: ./doc/config-reference/block-storage/drivers/ceph-rbd-volume-driver.xml93(para) msgid "" "To store and access your data, you can use the following storage systems:" msgstr "" #: ./doc/config-reference/block-storage/drivers/ceph-rbd-volume-driver.xml97(para) msgid "" "RADOS. Use as an object, default storage mechanism." msgstr "" #: ./doc/config-reference/block-storage/drivers/ceph-rbd-volume-driver.xml101(para) msgid "" "RBD. Use as a block device. The Linux kernel RBD (rados" " block device) driver allows striping a Linux block device over multiple " "distributed object store data objects. It is compatible with the kvm RBD " "image." msgstr "" #: ./doc/config-reference/block-storage/drivers/ceph-rbd-volume-driver.xml108(para) msgid "" "CephFS. Use as a file, POSIX-compliant file system." msgstr "" #: ./doc/config-reference/block-storage/drivers/ceph-rbd-volume-driver.xml112(para) msgid "" "Ceph exposes its distributed object store (RADOS). You can access it through" " the following interfaces:" msgstr "" #: ./doc/config-reference/block-storage/drivers/ceph-rbd-volume-driver.xml116(para) msgid "" "RADOS Gateway. Swift and Amazon-S3 compatible RESTful " "interface. See RADOS_Gateway for more " "information." msgstr "" #: ./doc/config-reference/block-storage/drivers/ceph-rbd-volume-driver.xml123(para) msgid "librados, and the related C/C++ bindings." msgstr "" #: ./doc/config-reference/block-storage/drivers/ceph-rbd-volume-driver.xml127(para) msgid "" "rbd and QEMU-RBD. Linux kernel and QEMU block devices " "that stripe data across multiple objects." msgstr "" #: ./doc/config-reference/block-storage/drivers/ceph-rbd-volume-driver.xml132(para) msgid "" "For detailed installation instructions and benchmarking information, see " "http://www.sebastien-han.fr/blog/2012/06/10/introducing-" "ceph-to-openstack/." msgstr "" #: ./doc/config-reference/block-storage/drivers/ceph-rbd-volume-driver.xml138(title) msgid "Driver options" msgstr "" #: ./doc/config-reference/block-storage/drivers/ceph-rbd-volume-driver.xml139(para) msgid "" "The following table contains the configuration options supported by the Ceph" " RADOS Block Device driver." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-lefthand-driver.xml4(title) msgid "HP / LeftHand SAN" msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-lefthand-driver.xml5(para) msgid "" "HP/LeftHand SANs are optimized for virtualized environments with VMware ESX " "& Microsoft Hyper-V, though the OpenStack integration provides " "additional support to various other virtualized environments, such as Xen, " "KVM, and OpenVZ, by exposing the volumes through ISCSI to connect to the " "instances." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-lefthand-driver.xml11(para) msgid "" "The HpSanISCSIDriver enables you to use a HP/Lefthand SAN that supports the " "Cliq interface. Every supported volume operation translates into a cliq call" " in the back-end." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-lefthand-driver.xml14(para) msgid "" "To use Cinder with HP/Lefthand SAN, you must set the following parameters in" " the cinder.conf file:" msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-lefthand-driver.xml19(para) msgid "" "Set " "volume_driver=cinder.volume.drivers.san.HpSanISCSIDriver." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-lefthand-driver.xml23(para) msgid "" "Set san_ip flag to the hostname or VIP of your " "Virtual Storage Appliance (VSA)." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-lefthand-driver.xml28(para) msgid "" "Set san_login and san_password" " to the user name and password of the ssh user with all necessary privileges" " on the appliance." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-lefthand-driver.xml34(para) msgid "" "Set san_ssh_port=16022. The default is 22. However, the default" " for the VSA is usually 16022." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-lefthand-driver.xml39(para) msgid "" "Set san_clustername to the name of the cluster where the " "associated volumes are created." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-lefthand-driver.xml44(para) msgid "The following optional parameters have the following default values:" msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-lefthand-driver.xml48(para) msgid "" "san_thin_provision=True. To disable thin provisioning, set to " "False." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-lefthand-driver.xml53(para) msgid "" "san_is_local=False. Typically, this parameter is set to " "False for this driver. To configure the cliq commands to " "run locally instead of over ssh, set this parameter to " "True." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-lefthand-driver.xml60(para) msgid "" "In addition to configuring the cinder-" "volume service, you must configure the VSA to function in an " "OpenStack environment." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-lefthand-driver.xml64(title) msgid "To configure the VSA" msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-lefthand-driver.xml66(para) msgid "" "Configure Chap on each of the nova-" "compute nodes." msgstr "" #: ./doc/config-reference/block-storage/drivers/hp-lefthand-driver.xml71(para) msgid "" "Add Server associations on the VSA with the associated Chap and initiator " "information. The name should correspond to the 'hostname' of the nova-compute node. For Xen, this is the hypervisor host " "name. To do this, use either Cliq or the Centralized Management Console." msgstr "" #: ./doc/config-reference/block-storage/drivers/windows-volume-driver.xml6(title) msgid "Windows" msgstr "" #: ./doc/config-reference/block-storage/drivers/windows-volume-driver.xml7(para) msgid "" "There is a volume back-end for Windows. Set the following in your " "cinder.conf, and use the options below to configure it." msgstr "" #: ./doc/config-reference/block-storage/drivers/lvm-volume-driver.xml6(title) msgid "LVM" msgstr "" #: ./doc/config-reference/block-storage/drivers/lvm-volume-driver.xml7(para) msgid "The default volume back-end uses local volumes managed by LVM." msgstr "" #: ./doc/config-reference/block-storage/drivers/lvm-volume-driver.xml8(para) msgid "" "This driver supports different transport protocols to attach volumes, " "currently ISCSI and ISER." msgstr "" #: ./doc/config-reference/block-storage/drivers/lvm-volume-driver.xml10(para) msgid "" "Set the following in your cinder.conf, and use the " "following options to configure for ISCSI transport:" msgstr "" #: ./doc/config-reference/block-storage/drivers/lvm-volume-driver.xml16(para) msgid "and for the ISER transport:" msgstr "" #: ./doc/config-reference/block-storage/drivers/nfs-volume-driver.xml4(title) msgid "NFS driver" msgstr "" #: ./doc/config-reference/block-storage/drivers/nfs-volume-driver.xml5(para) msgid "" "The Network File System (NFS) is a distributed file system protocol " "originally developed by Sun Microsystems in 1984. An NFS server " "exports one or more of its file systems, known as " "shares. An NFS client can mount these exported shares " "on its own file system. You can perform file actions on this mounted remote " "file system as if the file system were local." msgstr "" #: ./doc/config-reference/block-storage/drivers/nfs-volume-driver.xml13(title) msgid "How the NFS driver works" msgstr "" #: ./doc/config-reference/block-storage/drivers/nfs-volume-driver.xml14(para) msgid "" "The NFS driver, and other drivers based off of it, work quite differently " "than a traditional block storage driver." msgstr "" #: ./doc/config-reference/block-storage/drivers/nfs-volume-driver.xml17(para) msgid "" "The NFS driver does not actually allow an instance to access a storage " "device at the block level. Instead, files are created on an NFS share and " "mapped to instances, which emulates a block device. This works in a similar " "way to QEMU, which stores instances in the " "/var/lib/nova/instances directory." msgstr "" #: ./doc/config-reference/block-storage/drivers/nfs-volume-driver.xml26(title) msgid "Enable the NFS driver and related options" msgstr "" #: ./doc/config-reference/block-storage/drivers/nfs-volume-driver.xml27(para) msgid "" "To use Cinder with the NFS driver, first set the " "volume_driver in cinder.conf:" msgstr "" #: ./doc/config-reference/block-storage/drivers/nfs-volume-driver.xml31(para) msgid "The following table contains the options supported by the NFS driver." msgstr "" #: ./doc/config-reference/block-storage/drivers/nfs-volume-driver.xml37(title) msgid "How to use the NFS driver" msgstr "" #: ./doc/config-reference/block-storage/drivers/nfs-volume-driver.xml40(para) msgid "" "Access to one or more NFS servers. Creating an NFS server is outside the " "scope of this document. This example assumes access to the following NFS " "servers and mount points:" msgstr "" #: ./doc/config-reference/block-storage/drivers/nfs-volume-driver.xml46(literal) msgid "192.168.1.200:/storage" msgstr "" #: ./doc/config-reference/block-storage/drivers/nfs-volume-driver.xml49(literal) msgid "192.168.1.201:/storage" msgstr "" #: ./doc/config-reference/block-storage/drivers/nfs-volume-driver.xml52(literal) msgid "192.168.1.202:/storage" msgstr "" #: ./doc/config-reference/block-storage/drivers/nfs-volume-driver.xml55(para) msgid "" "This example demonstrates the use of with this driver with multiple NFS " "servers. Multiple servers are not required. One is usually enough." msgstr "" #: ./doc/config-reference/block-storage/drivers/nfs-volume-driver.xml60(para) msgid "" "Add your list of NFS servers to the file you specified with the " "nfs_shares_config option. For example, if the value of " "this option was set to /etc/cinder/shares.txt, then:" msgstr "" #: ./doc/config-reference/block-storage/drivers/nfs-volume-driver.xml70(para) msgid "" "Comments are allowed in this file. They begin with a #." msgstr "" #: ./doc/config-reference/block-storage/drivers/nfs-volume-driver.xml74(para) msgid "" "Configure the nfs_mount_point_base option. This is a " "directory where cinder-volume " "mounts all NFS shares stored in shares.txt. For this " "example, /var/lib/cinder/nfs is used. You can, of course," " use the default value of $state_path/mnt." msgstr "" #: ./doc/config-reference/block-storage/drivers/nfs-volume-driver.xml85(para) msgid "" "Start the cinder-volume service. " "/var/lib/cinder/nfs should now contain a directory for " "each NFS share specified in shares.txt. The name of each " "directory is a hashed name:" msgstr "" #: ./doc/config-reference/block-storage/drivers/nfs-volume-driver.xml97(para) msgid "You can now create volumes as you normally would:" msgstr "" #: ./doc/config-reference/block-storage/drivers/nfs-volume-driver.xml102(para) msgid "" "This volume can also be attached and deleted just like other volumes. " "However, snapshotting is not supported." msgstr "" #: ./doc/config-reference/block-storage/drivers/nfs-volume-driver.xml109(title) msgid "NFS driver notes" msgstr "" #: ./doc/config-reference/block-storage/drivers/nfs-volume-driver.xml112(para) msgid "" "cinder-volume manages the " "mounting of the NFS shares as well as volume creation on the shares. Keep " "this in mind when planning your OpenStack architecture. If you have one " "master NFS server, it might make sense to only have one cinder-volume service to handle all requests " "to that NFS server. However, if that single server is unable to handle all " "requests, more than one cinder-" "volume service is needed as well as potentially more than one " "NFS server." msgstr "" #: ./doc/config-reference/block-storage/drivers/nfs-volume-driver.xml128(para) msgid "" "Because data is stored in a file and not actually on a block storage device," " you might not see the same IO performance as you would with a traditional " "block storage driver. Please test accordingly." msgstr "" #: ./doc/config-reference/block-storage/drivers/nfs-volume-driver.xml135(para) msgid "" "Despite possible IO performance loss, having volume data stored in a file " "might be beneficial. For example, backing up volumes can be as easy as " "copying the volume files." msgstr "" #: ./doc/config-reference/block-storage/drivers/nfs-volume-driver.xml140(para) msgid "Regular IO flushing and syncing still stands." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-gpfs-volume-driver.xml4(title) msgid "IBM GPFS volume driver" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-gpfs-volume-driver.xml5(para) msgid "" "IBM General Parallel File System (GPFS) is a cluster file system that " "provides concurrent access to file systems from multiple nodes. The storage " "provided by these nodes can be direct attached, network attached, SAN " "attached, or a combination of these methods. GPFS provides many features " "beyond common data access, including data replication, policy based storage " "management, and space efficient file snapshot and clone operations." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-gpfs-volume-driver.xml14(title) msgid "How the GPFS driver works" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-gpfs-volume-driver.xml15(para) msgid "" "The GPFS driver enables the use of GPFS in a fashion similar to that of the " "NFS driver. With the GPFS driver, instances do not actually access a storage" " device at the block level. Instead, volume backing files are created in a " "GPFS file system and mapped to instances, which emulate a block device." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-gpfs-volume-driver.xml23(para) msgid "" "GPFS software must be installed and running on nodes where Block Storage and" " Compute services run in the OpenStack environment. A GPFS file system must " "also be created and mounted on these nodes before starting the cinder-volume service. The details of these GPFS specific steps " "are covered in GPFS: Concepts, Planning, and Installation " "Guide and GPFS: Administration and Programming " "Reference." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-gpfs-volume-driver.xml35(para) msgid "" "Optionally, the Image Service can be configured to store images on a GPFS " "file system. When a Block Storage volume is created from an image, if both " "image data and volume data reside in the same GPFS file system, the data " "from image file is moved efficiently to the volume file using copy-on-write " "optimization strategy." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-gpfs-volume-driver.xml43(title) msgid "Enable the GPFS driver" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-gpfs-volume-driver.xml44(para) msgid "" "To use the Block Storage Service with the GPFS driver, first set the " "volume_driver in cinder.conf:" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-gpfs-volume-driver.xml48(para) msgid "" "The following table contains the configuration options supported by the GPFS" " driver." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-gpfs-volume-driver.xml53(para) msgid "" "The gpfs_images_share_mode flag is only valid if the " "Image Service is configured to use GPFS with the " "gpfs_images_dir flag. When the value of this flag is " "copy_on_write, the paths specified by the " "gpfs_mount_point_base and " "gpfs_images_dir flags must both reside in the same GPFS " "file system and in the same GPFS file set." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-gpfs-volume-driver.xml66(title) msgid "Volume creation options" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-gpfs-volume-driver.xml67(para) msgid "" "It is possible to specify additional volume configuration options on a per-" "volume basis by specifying volume metadata. The volume is created using the " "specified options. Changing the metadata after the volume is created has no " "effect. The following table lists the volume creation options supported by " "the GPFS volume driver." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-gpfs-volume-driver.xml75(caption) msgid "Volume Create Options for GPFS Volume Drive" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-gpfs-volume-driver.xml79(th) msgid "Metadata Item Name" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-gpfs-volume-driver.xml85(literal) #: ./doc/config-reference/block-storage/drivers/ibm-gpfs-volume-driver.xml98(literal) #: ./doc/config-reference/block-storage/drivers/ibm-gpfs-volume-driver.xml99(literal) msgid "fstype" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-gpfs-volume-driver.xml88(literal) msgid "fstype=swap" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-gpfs-volume-driver.xml86(td) msgid "" "Specifies whether to create a file system or a swap area on the new volume. " "If is specified, the mkswap command is used to create a " "swap area. Otherwise the mkfs command is passed the specified file system " "type, for example ext3, ext4 or ntfs." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-gpfs-volume-driver.xml96(literal) msgid "fslabel" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-gpfs-volume-driver.xml97(td) msgid "" "Sets the file system label for the file system specified by " " option. This value is only used if is specified." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-gpfs-volume-driver.xml103(literal) msgid "data_pool_name" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-gpfs-volume-driver.xml105(para) msgid "" "Specifies the GPFS storage pool to which the volume is to be assigned. Note:" " The GPFS storage pool must already have been created." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-gpfs-volume-driver.xml111(literal) msgid "replicas" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-gpfs-volume-driver.xml113(para) msgid "" "Specifies how many copies of the volume file to create. Valid values are 1, " "2, and, for GPFS V3.5.0.7 and later, 3. This value cannot be greater than " "the value of the MaxDataReplicas attribute of the file " "system." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-gpfs-volume-driver.xml122(literal) msgid "dio" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-gpfs-volume-driver.xml124(para) msgid "" "Enables or disables the Direct I/O caching policy for the volume file. Valid" " values are yes and no." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-gpfs-volume-driver.xml130(literal) msgid "write_affinity_depth" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-gpfs-volume-driver.xml132(para) msgid "" "Specifies the allocation policy to be used for the volume file. Note: This " "option only works if allow-write-affinity is set for the " "GPFS data pool." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-gpfs-volume-driver.xml139(literal) msgid "block_group_factor" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-gpfs-volume-driver.xml141(para) msgid "" "Specifies how many blocks are laid out sequentially in the volume file to " "behave as a single large block. Note: This option only works if allow-write-affinity is set for the GPFS data pool." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-gpfs-volume-driver.xml149(literal) msgid "write_affinity_failure_group" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-gpfs-volume-driver.xml151(para) msgid "" "Specifies the range of nodes (in GPFS shared nothing architecture) where " "replicas of blocks in the volume file are to be written. See " "GPFS: Administration and Programming Reference for " "more details on this option." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-gpfs-volume-driver.xml162(title) msgid "Example: Volume creation options" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-gpfs-volume-driver.xml163(para) msgid "" "This example shows the creation of a 50GB volume with an ext4 file system " "labeled newfsand direct IO enabled:" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-gpfs-volume-driver.xml171(title) msgid "Operational notes for GPFS driver" msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-gpfs-volume-driver.xml174(para) msgid "" "Volume snapshots are implemented using the GPFS file clone feature. Whenever" " a new snapshot is created, the snapshot file is efficiently created as a " "read-only clone parent of the volume, and the volume file uses copy-on-write" " optimization strategy to minimize data movement." msgstr "" #: ./doc/config-reference/block-storage/drivers/ibm-gpfs-volume-driver.xml180(para) msgid "" "Similarly when a new volume is created from a snapshot or from an existing " "volume, the same approach is taken. The same approach is also used when a " "new volume is created from a Glance image, if the source image is in raw " "format, and gpfs_images_share_mode is set to " "copy_on_write." msgstr "" #: ./doc/config-reference/block-storage/backup/tsm-backup-driver.xml6(title) msgid "IBM Tivoli Storage Manager backup driver" msgstr "" #: ./doc/config-reference/block-storage/backup/tsm-backup-driver.xml7(para) msgid "" "The IBM Tivoli Storage Manager (TSM) backup driver enables performing volume" " backups to a TSM server." msgstr "" #: ./doc/config-reference/block-storage/backup/tsm-backup-driver.xml10(para) msgid "" "The TSM client should be installed and configured on the machine running the" " cinder-backup service. See the " "IBM Tivoli Storage Manager Backup-Archive Client Installation and" " User's Guide for details on installing the TSM client." msgstr "" #: ./doc/config-reference/block-storage/backup/tsm-backup-driver.xml17(para) msgid "" "To enable the IBM TSM backup driver, include the following option in " "cinder.conf:" msgstr "" #: ./doc/config-reference/block-storage/backup/tsm-backup-driver.xml20(para) msgid "" "The following configuration options are available for the TSM backup driver." msgstr "" #: ./doc/config-reference/block-storage/backup/tsm-backup-driver.xml23(para) msgid "This example shows the default options for the TSM backup driver." msgstr "" #: ./doc/config-reference/block-storage/backup/ceph-backup-driver.xml11(title) msgid "Ceph backup driver" msgstr "" #: ./doc/config-reference/block-storage/backup/ceph-backup-driver.xml12(para) msgid "" "The Ceph backup driver backs up volumes of any type to a Ceph back-end " "store. The driver can also detect whether the volume to be backed up is a " "Ceph RBD volume, and if so, it tries to perform incremental and differential" " backups." msgstr "" #: ./doc/config-reference/block-storage/backup/ceph-backup-driver.xml17(para) msgid "" "For source Ceph RBD volumes, you can perform backups within the same Ceph " "pool (not recommended) and backups between different Ceph pools and between " "different Ceph clusters." msgstr "" #: ./doc/config-reference/block-storage/backup/ceph-backup-driver.xml21(para) msgid "" "At the time of writing, differential backup support in Ceph/librbd was quite" " new. This driver attempts a differential backup in the first instance. If " "the differential backup fails, the driver falls back to full backup/copy." msgstr "" #: ./doc/config-reference/block-storage/backup/ceph-backup-driver.xml26(para) msgid "" "If incremental backups are used, multiple backups of the same volume are " "stored as snapshots so that minimal space is consumed in the backup store. " "It takes far less time to restore a volume than to take a full copy." msgstr "" #: ./doc/config-reference/block-storage/backup/ceph-backup-driver.xml32(para) msgid "Block Storage Service enables you to:" msgstr "" #: ./doc/config-reference/block-storage/backup/ceph-backup-driver.xml35(para) msgid "Restore to a new volume, which is the default and recommended action." msgstr "" #: ./doc/config-reference/block-storage/backup/ceph-backup-driver.xml40(para) msgid "" "Restore to the original volume from which the backup was taken. The restore " "action takes a full copy because this is the safest action." msgstr "" #: ./doc/config-reference/block-storage/backup/ceph-backup-driver.xml48(para) msgid "" "To enable the Ceph backup driver, include the following option in the " "cinder.conf file:" msgstr "" #: ./doc/config-reference/block-storage/backup/ceph-backup-driver.xml52(para) msgid "" "The following configuration options are available for the Ceph backup " "driver." msgstr "" #: ./doc/config-reference/block-storage/backup/ceph-backup-driver.xml56(para) msgid "This example shows the default options for the Ceph backup driver." msgstr "" #: ./doc/config-reference/block-storage/backup/swift-backup-driver.xml5(title) msgid "Swift backup driver" msgstr "" #: ./doc/config-reference/block-storage/backup/swift-backup-driver.xml6(para) msgid "" "The backup driver for Swift back-end performs a volume backup to a Swift " "object storage system." msgstr "" #: ./doc/config-reference/block-storage/backup/swift-backup-driver.xml8(para) msgid "" "To enable the Swift backup driver, include the following option in the " "cinder.conf file:" msgstr "" #: ./doc/config-reference/block-storage/backup/swift-backup-driver.xml12(para) msgid "" "The following configuration options are available for the Swift back-end " "backup driver." msgstr "" #: ./doc/config-reference/block-storage/backup/swift-backup-driver.xml16(para) msgid "" "This example shows the default options for the Swift back-end backup driver." msgstr "" #: ./doc/config-reference/image/section_glance-property-protection.xml6(title) msgid "Image property protection" msgstr "" #: ./doc/config-reference/image/section_glance-property-protection.xml7(para) msgid "" "There are currently two types of properties in the Image Service: \"core " "properties,\" which are defined by the system, and \"additional " "properties,\" which are arbitrary key/value pairs that can be set on an " "image." msgstr "" #: ./doc/config-reference/image/section_glance-property-protection.xml11(para) msgid "" "With the Havana release, any such property can be protected through " "configuration. When you put protections on a property, it limits the users " "who can perform CRUD operations on the property based on their user role. " "The use case is to enable the cloud provider to maintain extra properties on" " images so typically this would be an administrator who has access to " "protected properties, managed with policy.json. The " "extra property could be licensing information or billing information, for " "example." msgstr "" #: ./doc/config-reference/image/section_glance-property-protection.xml20(para) msgid "" "Properties that don't have protections defined for them will act as they do " "now: the administrator can control core properties, with the image owner " "having control over additional properties." msgstr "" #: ./doc/config-reference/image/section_glance-property-protection.xml23(para) msgid "" "Property protection can be set in /etc/glance/property-" "protections.conf, using roles found in " "policy.json." msgstr "" #: ./doc/config-reference/object-storage/section_configure_s3.xml6(title) msgid "Configure Object Storage with the S3 API" msgstr "" #: ./doc/config-reference/object-storage/section_configure_s3.xml7(para) msgid "" "The Swift3 middleware emulates the S3 REST API on top of Object Storage." msgstr "" #: ./doc/config-reference/object-storage/section_configure_s3.xml9(para) msgid "The following operations are currently supported:" msgstr "" #: ./doc/config-reference/object-storage/section_configure_s3.xml12(para) msgid "GET Service" msgstr "" #: ./doc/config-reference/object-storage/section_configure_s3.xml15(para) msgid "DELETE Bucket" msgstr "" #: ./doc/config-reference/object-storage/section_configure_s3.xml18(para) msgid "GET Bucket (List Objects)" msgstr "" #: ./doc/config-reference/object-storage/section_configure_s3.xml21(para) msgid "PUT Bucket" msgstr "" #: ./doc/config-reference/object-storage/section_configure_s3.xml24(para) msgid "DELETE Object" msgstr "" #: ./doc/config-reference/object-storage/section_configure_s3.xml27(para) msgid "GET Object" msgstr "" #: ./doc/config-reference/object-storage/section_configure_s3.xml30(para) msgid "HEAD Object" msgstr "" #: ./doc/config-reference/object-storage/section_configure_s3.xml33(para) msgid "PUT Object" msgstr "" #: ./doc/config-reference/object-storage/section_configure_s3.xml36(para) msgid "PUT Object (Copy)" msgstr "" #: ./doc/config-reference/object-storage/section_configure_s3.xml39(para) msgid "" "To use this middleware, first download the latest version from its " "repository to your proxy server(s)." msgstr "" #: ./doc/config-reference/object-storage/section_configure_s3.xml43(para) msgid "" "Optional: To use this middleware with Swift 1.7.0 and previous versions, you" " must use the v1.7 tag of the fujita/swift3 repository. Clone the " "repository, as shown previously, and run this command:" msgstr "" #: ./doc/config-reference/object-storage/section_configure_s3.xml48(para) msgid "Then, install it using standard python mechanisms, such as:" msgstr "" #: ./doc/config-reference/object-storage/section_configure_s3.xml51(para) msgid "" "Alternatively, if you have configured the Ubuntu Cloud Archive, you may use:" " " msgstr "" #: ./doc/config-reference/object-storage/section_configure_s3.xml54(para) msgid "" "To add this middleware to your configuration, add the swift3 middleware in " "front of the auth middleware, and before any other middleware that look at " "swift requests (like rate limiting)." msgstr "" #: ./doc/config-reference/object-storage/section_configure_s3.xml58(para) msgid "" "Ensure that your proxy-server.conf file contains swift3 in the pipeline and " "the [filter:swift3] section, as shown below:" msgstr "" #: ./doc/config-reference/object-storage/section_configure_s3.xml68(para) msgid "" "Next, configure the tool that you use to connect to the S3 API. For S3curl, " "for example, you must add your host IP information by adding your host IP to" " the @endpoints array (line 33 in s3curl.pl):" msgstr "" #: ./doc/config-reference/object-storage/section_configure_s3.xml73(para) msgid "Now you can send commands to the endpoint, such as:" msgstr "" #: ./doc/config-reference/object-storage/section_configure_s3.xml77(para) msgid "" "To set up your client, the access key will be the concatenation of the " "account and user strings that should look like test:tester, and the secret " "access key is the account password. The host should also point to the Swift " "storage node's hostname. It also will have to use the old-style calling " "format, and not the hostname-based container format. Here is an example " "client setup using the Python boto library on a locally installed all-in-one" " Swift installation." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml6(title) msgid "Configure Object Storage features" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml8(title) msgid "Object Storage zones" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml9(para) msgid "" "In OpenStack Object Storage, data is placed across different tiers of " "failure domains. First, data is spread across regions, then zones, then " "servers, and finally across drives. Data is placed to get the highest " "failure domain isolation. If you deploy multiple regions, the Object Storage" " service places the data across the regions. Within a region, each replica " "of the data should be stored in unique zones, if possible. If there is only " "one zone, data should be placed on different servers. And if there is only " "one server, data should be placed on different drives." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml20(para) msgid "" "Regions are widely separated installations with a high-latency or otherwise " "constrained network link between them. Zones are arbitrarily assigned, and " "it is up to the administrator of the Object Storage cluster to choose an " "isolation level and attempt to maintain the isolation level through " "appropriate zone assignment. For example, a zone may be defined as a rack " "with a single power source. Or a zone may be a DC room with a common utility" " provider. Servers are identified by a unique IP/port. Drives are locally " "attached storage volumes identified by mount point." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml31(para) msgid "" "In small clusters (five nodes or fewer), everything is normally in a single " "zone. Larger Object Storage deployments may assign zone designations " "differently; for example, an entire cabinet or rack of servers may be " "designated as a single zone to maintain replica availability if the cabinet " "becomes unavailable (for example, due to failure of the top of rack switches" " or a dedicated circuit). In very large deployments, such as service " "provider level deployments, each zone might have an entirely autonomous " "switching and power infrastructure, so that even the loss of an electrical " "circuit or switching aggregator would result in the loss of a single replica" " at most." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml45(title) msgid "Rackspace zone recommendations" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml46(para) msgid "" "For ease of maintenance on OpenStack Object Storage, Rackspace recommends " "that you set up at least five nodes. Each node is assigned its own zone (for" " a total of five zones), which gives you host level redundancy. This enables" " you to take down a single zone for maintenance and still guarantee object " "availability in the event that another zone fails during your maintenance." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml54(para) msgid "" "You could keep each server in its own cabinet to achieve cabinet level " "isolation, but you may wish to wait until your swift service is better " "established before developing cabinet-level isolation. OpenStack Object " "Storage is flexible; if you later decide to change the isolation level, you " "can take down one zone at a time and move them to appropriate new homes." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml65(title) msgid "RAID controller configuration" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml66(para) msgid "" "OpenStack Object Storage does not require RAID. In fact, most RAID " "configurations cause significant performance degradation. The main reason " "for using a RAID controller is the battery-backed cache. It is very " "important for data integrity reasons that when the operating system confirms" " a write has been committed that the write has actually been committed to a " "persistent location. Most disks lie about hardware commits by default, " "instead writing to a faster write cache for performance reasons. In most " "cases, that write cache exists only in non-persistent memory. In the case of" " a loss of power, this data may never actually get committed to disk, " "resulting in discrepancies that the underlying file system must handle." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml79(para) msgid "" "OpenStack Object Storage works best on the XFS file system, and this " "document assumes that the hardware being used is configured appropriately to" " be mounted with the option.   For more information, refer " "to the XFS FAQ: http://xfs.org/index.php/XFS_FAQ" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml87(para) msgid "" "To get the most out of your hardware, it is essential that every disk used " "in OpenStack Object Storage is configured as a standalone, individual RAID 0" " disk; in the case of 6 disks, you would have six RAID 0s or one JBOD. Some " "RAID controllers do not support JBOD or do not support battery backed cache " "with JBOD. To ensure the integrity of your data, you must ensure that the " "individual drive caches are disabled and the battery backed cache in your " "RAID card is configured and used. Failure to configure the controller " "properly in this case puts data at risk in the case of sudden loss of power." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml99(para) msgid "" "You can also use hybrid drives or similar options for battery backed up " "cache configurations without a RAID controller." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml105(title) msgid "Throttle resources through rate limits" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml106(para) msgid "" "Rate limiting in OpenStack Object Storage is implemented as a pluggable " "middleware that you configure on the proxy server. Rate limiting is " "performed on requests that result in database writes to the account and " "container SQLite databases. It uses memcached and is dependent on the proxy " "servers having highly synchronized time. The rate limits are limited by the " "accuracy of the proxy server clocks." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml115(title) msgid "Configure rate limiting" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml116(para) msgid "" "All configuration is optional. If no account or container limits are " "provided, no rate limiting occurs. Available configuration options include:" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml122(para) msgid "" "The container rate limits are linearly interpolated from the values given. A" " sample container rate limiting could be:" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml125(para) msgid "container_ratelimit_100 = 100" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml126(para) msgid "container_ratelimit_200 = 50" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml127(para) msgid "container_ratelimit_500 = 20" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml128(para) msgid "This would result in:" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml130(caption) msgid "Values for Rate Limiting with Sample Configuration Settings" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml134(td) msgid "Container Size" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml135(td) msgid "Rate Limit" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml138(td) msgid "0-99" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml139(td) msgid "No limiting" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml142(td) #: ./doc/config-reference/object-storage/section_object-storage-features.xml143(td) msgid "100" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml147(td) msgid "150" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml148(td) msgid "75" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml151(td) msgid "500" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml152(td) #: ./doc/config-reference/object-storage/section_object-storage-features.xml156(td) msgid "20" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml155(td) msgid "1000" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml163(title) msgid "Health check" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml164(para) msgid "" "Provides an easy way to monitor whether the swift proxy server is alive. If " "you access the proxy with the path /healthcheck, it " "respond OK in the response body, which monitoring tools " "can use." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml174(title) msgid "Domain remap" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml175(para) msgid "" "Middleware that translates container and account parts of a domain to path " "parameters that the proxy server understands." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml183(title) msgid "CNAME lookup" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml184(para) msgid "" "Middleware that translates an unknown domain in the host header to something" " that ends with the configured storage_domain by looking up the given " "domain's CNAME record in DNS." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml194(title) msgid "Temporary URL" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml195(para) msgid "" "Allows the creation of URLs to provide temporary access to objects. For " "example, a website may wish to provide a link to download a large object in " "Swift, but the Swift account has no public access. The website can generate " "a URL that provides GET access for a limited time to the resource. When the " "web browser user clicks on the link, the browser downloads the object " "directly from Swift, eliminating the need for the website to act as a proxy " "for the request. If the user shares the link with all his friends, or " "accidentally posts it on a forum, the direct access is limited to the " "expiration time set when the website created the link." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml210(literal) msgid "temp_url_sig" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml212(para) msgid "A cryptographic signature" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml216(literal) msgid "temp_url_expires" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml218(para) msgid "An expiration date, in Unix time." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml207(para) msgid "" "A temporary URL is the typical URL associated with an object, with two " "additional query parameters:" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml222(para) msgid "An example of a temporary URL:" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml228(para) msgid "" "To create temporary URLs, first set the X-Account-Meta-Temp-URL-" "Key header on your Swift account to an arbitrary string. This " "string serves as a secret key. For example, to set a key of " "b3968d0207b54ece87cccc06515a89d4 using the " " command-line tool:" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml235(replaceable) msgid "b3968d0207b54ece87cccc06515a89d4" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml236(para) msgid "Next, generate an HMAC-SHA1 (RFC 2104) signature to specify:" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml240(para) msgid "" "Which HTTP method to allow (typically GET or " "PUT)" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml245(para) msgid "The expiry date as a Unix timestamp" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml248(para) msgid "the full path to the object" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml251(para) msgid "" "The secret key set as the X-Account-Meta-Temp-URL-Key" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml255(para) msgid "" "Here is code generating the signature for a GET for 24 hours on " "/v1/AUTH_account/container/object:" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml279(para) msgid "" "Changing the X-Account-Meta-Temp-URL-Key invalidates any " "previously generated temporary URLs within 60 seconds (the memcache time for" " the key). Swift supports up to two keys, specified by X-Account-" "Meta-Temp-URL-Key and X-Account-Meta-Temp-URL-" "Key-2. Signatures are checked against both keys, if present. This " "is to allow for key rotation without invalidating all existing temporary " "URLs." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml270(para) msgid "" "Any alteration of the resource path or query arguments results in a " "401Unauthorized error. " "Similarly, a PUT where GET was the allowed method returns a " "401. HEAD is allowed if GET or PUT is allowed. Using " "this in combination with browser form post translation middleware could also" " allow direct-from-browser uploads to specific locations in Swift. Note that" " " msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml291(para) msgid "" "Swift includes a script called that generates the query " "parameters automatically:" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml298(para) msgid "" "Because this command only returns the path, you must prefix the Swift " "storage host name (for example, https://swift-" "cluster.example.com)." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml301(para) msgid "" "With GET Temporary URLs, a Content-Disposition header is " "set on the response so that browsers interpret this as a file attachment to " "be saved. The file name chosen is based on the object name, but you can " "override this with a filename query parameter. The " "following example specifies a filename of My Test " "File.pdf:" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml324(emphasis) msgid "tempurl" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml313(para) msgid "" "To enable Temporary URL functionality, edit /etc/swift/proxy-" "server.conf to add tempurl to the " "pipeline variable defined in the " "[pipeline:main] section. The tempurl " "entry should appear immediately before the authentication filters in the " "pipeline, such as authtoken, tempauth " "or keystoneauth. For example:" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml331(title) msgid "Name check filter" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml332(para) msgid "" "Name Check is a filter that disallows any paths that contain defined " "forbidden characters or that exceed a defined length." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml340(title) msgid "Constraints" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml341(para) msgid "" "To change the OpenStack Object Storage internal limits, update the values in" " the swift-constraints section in the " "swift.conf file. Use caution when you update these " "values because they affect the performance in the entire cluster." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml352(title) msgid "Cluster health" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml353(para) msgid "" "Use the tool to measure overall cluster health. This tool " "checks if a set of deliberately distributed containers and objects are " "currently in their proper places within the cluster. For instance, a common " "deployment has three replicas of each object. The health of that object can " "be measured by checking if each replica is in its proper place. If only 2 of" " the 3 is in place the object’s health can be said to be at 66.66%, where " "100% would be perfect. A single object’s health, especially an older object," " usually reflects the health of that entire partition the object is in. If " "you make enough objects on a distinct percentage of the partitions in the " "cluster,you get a good estimate of the overall cluster health. In practice, " "about 1% partition coverage seems to balance well between accuracy and the " "amount of time it takes to gather results. The first thing that needs to be " "done to provide this health value is create a new account solely for this " "usage. Next, you need to place the containers and objects throughout the " "system so that they are on distinct partitions. The swift-dispersion-" "populate tool does this by making up random container and object names until" " they fall on distinct partitions. Last, and repeatedly for the life of the " "cluster, you must run the tool to check the health of each " "of these containers and objects. These tools need direct access to the " "entire cluster and to the ring files (installing them on a proxy server " "suffices). The and commands both use the " "same configuration file, /etc/swift/dispersion.conf. " "Example dispersion.conf file:" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml393(para) msgid "" "There are also configuration options for specifying the dispersion coverage," " which defaults to 1%, retries, concurrency, and so on. However, the " "defaults are usually fine. Once the configuration is in place, run " " to populate the containers and objects throughout the " "cluster. Now that those containers and objects are in place, you can run " " to get a dispersion report, or the overall health of the " "cluster. Here is an example of a cluster in perfect health:" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml413(para) msgid "" "Now, deliberately double the weight of a device in the object ring (with " "replication turned off) and re-run the dispersion report to show what impact" " that has:" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml429(para) msgid "" "You can see the health of the objects in the cluster has gone down " "significantly. Of course, this test environment has just four devices, in a " "production environment with many devices the impact of one device change is " "much less. Next, run the replicators to get everything put back into place " "and then rerun the dispersion report:" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml446(para) msgid "" "Alternatively, the dispersion report can also be output in json format. This" " allows it to be more easily consumed by third party utilities:" msgstr "" #. Usage documented in #. http://docs.openstack.org/developer/swift/overview_large_objects.html #: ./doc/config-reference/object-storage/section_object-storage-features.xml460(title) msgid "Static Large Object (SLO) support" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml461(para) msgid "" "This feature is very similar to Dynamic Large Object (DLO) support in that " "it enables the user to upload many objects concurrently and afterwards " "download them as a single object. It is different in that it does not rely " "on eventually consistent container listings to do so. Instead, a user " "defined manifest of the object segments is used." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml473(title) msgid "Container quotas" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml474(para) msgid "" "The container_quotas middleware implements simple quotas that can be imposed" " on swift containers by a user with the ability to set container metadata, " "most likely the account administrator. This can be useful for limiting the " "scope of containers that are delegated to non-admin users, exposed to " "formpost uploads, or just as a self-imposed sanity check." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml481(para) msgid "" "Any object PUT operations that exceed these quotas return a 413 response " "(request entity too large) with a descriptive body." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml484(para) msgid "" "Quotas are subject to several limitations: eventual consistency, the " "timeliness of the cached container_info (60 second ttl by default), and it " "is unable to reject chunked transfer uploads that exceed the quota (though " "once the quota is exceeded, new chunked transfers are refused)." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml490(para) msgid "" "Set quotas by adding meta values to the container. These values are " "validated when you set them:" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml494(para) msgid "X-Container-Meta-Quota-Bytes: Maximum size of the container, in bytes." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml498(para) msgid "X-Container-Meta-Quota-Count: Maximum object count of the container." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml507(title) msgid "Account quotas" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml508(para) msgid "" "The x-account-meta-quota-bytes metadata entry must be" " requests (PUT, POST) if a given account quota (in bytes) is exceeded while " "DELETE requests are still allowed." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml512(para) msgid "" "The x-account-meta-quota-bytes metadata entry must be set to store and " "enable the quota. Write requests to this metadata entry are only permitted " "for resellers. There is no account quota limitation on a reseller account " "even if x-account-meta-quota-bytes is set." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml517(para) msgid "" "Any object PUT operations that exceed the quota return a 413 response " "(request entity too large) with a descriptive body." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml520(para) msgid "" "The following command uses an admin account that own the Reseller role to " "set a quota on the test account:" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml524(para) msgid "Here is the stat listing of an account where quota has been set:" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml534(para) msgid "This command removes the account quota:" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml538(title) msgid "Bulk delete" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml539(para) msgid "" "Use bulk-delete to delete multiple files from an account with a single " "request. Responds to DELETE requests with a header 'X-Bulk-Delete: " "true_value'. The body of the DELETE request is a new line separated list of " "files to delete. The files listed must be URL encoded and in the form:" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml548(para) msgid "" "If all files are successfully deleted (or did not exist), the operation " "returns HTTPOk. If any files failed to delete, the operation returns " "HTTPBadGateway. In both cases the response body is a JSON dictionary that " "shows the number of files that were successfully deleted or not found. The " "files that failed are listed." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml560(title) msgid "Drive audit" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml561(para) msgid "" "The configuration items reference a " "script that can be run by using to watch for bad drives. If" " errors are detected, it unmounts the bad drive, so that OpenStack Object " "Storage can work around it. It takes the following options:" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml572(title) msgid "Form post" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml573(para) msgid "" "Middleware that provides the ability to upload objects to a cluster using an" " HTML form POST. The format of the form is:" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml588(para) msgid "" "The swift-url is the URL to the Swift destination, such " "as: https://swift-" "cluster.example.com/v1/AUTH_account/container/object_prefix The name " "of each file uploaded is appended to the specified swift-" "url. So, you can upload directly to the root of container with a " "url like: https://swift-" "cluster.example.com/v1/AUTH_account/container/ Optionally, you can " "include an object prefix to better separate different users’ uploads, such " "as: https://swift-" "cluster.example.com/v1/AUTH_account/container/object_prefix" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml600(para) msgid "" "The form method must be POST and the enctype must be set as " "multipart/form-data." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml603(para) msgid "" "The redirect attribute is the URL to redirect the browser to after the " "upload completes. The URL has status and message query parameters added to " "it, indicating the HTTP status code for the upload (2xx is success) and a " "possible message for further information if there was an error (such as " "“max_file_size exceeded”)." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml610(para) msgid "" "The max_file_size attribute must be included and " "indicates the largest single file upload that can be done, in bytes." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml613(para) msgid "" "The max_file_count attribute must be included and " "indicates the maximum number of files that can be uploaded with the form. " "Include additional <![CDATA[<input type=\"file\" " "name=\"filexx\"/>]]> attributes if desired." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml619(para) msgid "" "The expires attribute is the Unix timestamp before which the form must be " "submitted before it is invalidated." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml622(para) msgid "" "The signature attribute is the HMAC-SHA1 signature of the form. This sample " "Python code shows how to compute the signature:" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml639(para) msgid "" "The key is the value of the X-Account-Meta-Temp-URL-Key " "header on the account." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml642(para) msgid "" "Be certain to use the full path, from the /v1/ onward." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml644(para) msgid "" "The command line tool may be used (mostly just when " "testing) to compute expires and signature." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml648(para) msgid "" "The file attributes must appear after the other attributes to be processed " "correctly. If attributes come after the file, they are not sent with the " "sub-request because on the server side, all attributes in the file cannot be" " parsed unless the whole file is read into memory and the server does not " "have enough memory to service these requests. So, attributes that follow the" " file are ignored." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml661(title) msgid "Static web sites" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-features.xml662(para) msgid "" "When configured, this middleware serves container data as a static web site " "with index file and error file resolution and optional file listings. This " "mode is normally only active for anonymous requests." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-cors.xml6(title) msgid "Cross-origin resource sharing" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-cors.xml7(para) msgid "" "Cross-Origin Resource Sharing (CORS) is a mechanisim to allow code running " "in a browser (Javascript for example) to make requests to a domain other " "then the one from where it originated. Swift supports CORS requests to " "containers and objects within the containers using metadata held on the " "container." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-cors.xml13(para) msgid "" "In addition to the metadata on containers, the " "cors_allow_origin flag in proxy-" "server.conf may be used to set a list of hosts that are included " "with any CORS request by default." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-general-service-conf.xml6(title) msgid "Object Storage general service configuration" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-general-service-conf.xml7(para) msgid "" "Most Object Storage services fall into two categories, Object Storage's wsgi" " servers and background daemons." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-general-service-conf.xml11(para) msgid "" "Object Storage uses paste.deploy to manage server configurations. Read more " "at http://pythonpaste.org/deploy/." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-general-service-conf.xml14(para) msgid "" "Default configuration options are set in the `[DEFAULT]` section, and any " "options specified there can be overridden in any of the other sections when " "the syntax set option_name = value is in place." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-general-service-conf.xml19(para) msgid "" "Configuration for servers and daemons can be expressed together in the same " "file for each type of server, or separately. If a required section for the " "service trying to start is missing there will be an error. The sections not " "used by the service are ignored." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-general-service-conf.xml25(para) msgid "" "Consider the example of an object storage node. By convention configuration " "for the object-server, object-updater, object-replicator, and object-auditor" " exist in a single file /etc/swift/object-server.conf:" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-general-service-conf.xml47(para) msgid "" "Object Storage services expect a configuration path as the first argument:" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-general-service-conf.xml55(para) msgid "" "If you omit the object-auditor section this file can not be used as the " "configuration path when starting the daemon:" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-general-service-conf.xml63(para) msgid "" "If the configuration path is a directory instead of a file all of the files " "in the directory with the file extension \".conf\" will be combined to " "generate the configuration object which is delivered to the Object Storage " "service. This is referred to generally as \"directory based configuration\"." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-general-service-conf.xml70(para) msgid "" "Directory based configuration leverages ConfigParser's native multi-file " "support. Files ending in \".conf\" in the given directory are parsed in " "lexicographical order. File names starting with '.' are ignored. A mixture " "of file and directory configuration paths is not supported - if the " "configuration path is a file, only that file will be parsed." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-general-service-conf.xml78(para) msgid "" "The Object Storage service management tool swift-init " "has adopted the convention of looking for " "/etc/swift/{type}-server.conf.d/ if the file " "/etc/swift/{type}-server.conf file does not exist." msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-general-service-conf.xml86(para) msgid "" "When using directory based configuration, if the same option under the same " "section appears more than once in different files, the last value parsed is " "said to override previous occurrences. You can ensure proper override " "precedence by prefixing the files in the configuration directory with " "numerical values, as in the following example file layout:" msgstr "" #: ./doc/config-reference/object-storage/section_object-storage-general-service-conf.xml105(para) msgid "" "You can inspect the resulting combined configuration object using the " " command line tool." msgstr "" #: ./doc/config-reference/identity/section_keystone-token-binding.xml6(title) msgid "Configure the Identity Service for token binding" msgstr "" #: ./doc/config-reference/identity/section_keystone-token-binding.xml7(para) msgid "" "Token binding refers to the practice of embedding information from external " "authentication providers (like a company's Kerberos server) inside the token" " such that a client may enforce that the token only be used in conjunction " "with that specified authentication. This is an additional security mechanism" " as it means that if a token is stolen it will not be usable without also " "providing the external authentication." msgstr "" #: ./doc/config-reference/identity/section_keystone-token-binding.xml14(para) msgid "" "To activate token binding you must specify the types of authentication that " "token binding should be used for in keystone.conf: " " Currently only kerberos is supported." msgstr "" #: ./doc/config-reference/identity/section_keystone-token-binding.xml25(para) msgid "disabled disable token bind checking" msgstr "" #: ./doc/config-reference/identity/section_keystone-token-binding.xml29(para) msgid "" "permissive enable bind checking, if a token is bound to a" " mechanism that is unknown to the server then ignore it. This is the " "default." msgstr "" #: ./doc/config-reference/identity/section_keystone-token-binding.xml34(para) msgid "" "strict enable bind checking, if a token is bound to a " "mechanism that is unknown to the server then this token should be rejected." msgstr "" #: ./doc/config-reference/identity/section_keystone-token-binding.xml39(para) msgid "" "required enable bind checking and require that at least 1" " bind mechanism is used for tokens." msgstr "" #: ./doc/config-reference/identity/section_keystone-token-binding.xml44(para) msgid "" "named enable bind checking and require that the specified" " authentication mechanism is used: " msgstr "" #: ./doc/config-reference/identity/section_keystone-token-binding.xml51(para) msgid "" "Do not set enforce_token_bind = named as there is not an " "authentication mechanism called named." msgstr "" #: ./doc/config-reference/identity/section_keystone-token-binding.xml20(para) msgid "" "To enforce checking of token binding the " "enforce_token_bind parameter should be set to one of the " "following modes: " msgstr "" #. Put one translator per line, in the form of NAME , YEAR1, YEAR2 #: ./doc/config-reference/identity/section_keystone-token-binding.xml0(None) msgid "translator-credits" msgstr "ترجمه اعتبارات , مترجم-کریڈٹس"