System for quickly installing an OpenStack cloud from upstream git for testing and development.
Go to file
Mate Lakat 68ac03c7f2 xenapi: separate disk for cinder volumes
Some kernels have lockup issues while serving cinder volumes from
file-loopback-lvm kind of cinder setup. This patch provides a way to
create an additional drive to the OpenStack VM, and use that to store
cinder volumes, thus eliminating the lockup issue. It will help when
testing XenServer. Now, you can specify devices for stack-volumes
through:

    VOLUME_BACKING_DEVICE

In case you are using CINDER_MULTI_LVM_BACKEND, you can use

    VOLUME_BACKING_DEVICE2

as well.

Xenserver:
Should you whish to use a 10 gig disk backend for your cinder volumes,
specify:

    XEN_XVDB_SIZE_GB=10
    VOLUME_BACKING_DEVICE=/dev/xvdb

Citrix is using this approach on its internal CI system to run tests
against OpenStack. This is a workaround for this bug:

  https://bugs.launchpad.net/cinder/+bug/1023755

Related to blueprint xenapi-devstack-cleanup

Change-Id: Iee633d2704185bfbf9234882654c47b850fa168a
2013-06-19 13:56:33 +01:00
exercises exercise euca.sh check the volume just created 2013-06-16 10:28:23 +08:00
extras.d Add basic uec image preparation to tempest 2013-01-15 14:23:07 +01:00
files Merge "Update the ou name of Group to be the new default of UserGroups" 2013-06-19 11:08:02 +00:00
lib xenapi: separate disk for cinder volumes 2013-06-19 13:56:33 +01:00
samples add command for Add icmp tcp/22 to default security group 2012-09-09 11:19:58 +08:00
tests Support MultiStrOpt options in configuration file. 2013-03-06 09:42:14 +08:00
tools xenapi: separate disk for cinder volumes 2013-06-19 13:56:33 +01:00
.gitignore Ignore autogenerated .prereqs. 2013-03-01 10:31:29 +01:00
.gitreview Add .gitreview config file for gerrit. 2011-11-16 11:24:49 -08:00
.mailmap Update my mailmap 2013-06-05 15:31:09 -07:00
AUTHORS Add myself to AUTHORS and add my email addresses to .mailmap 2013-05-25 00:43:44 +08:00
clean.sh Merge "On unstack, cleanup LVM setup for Cinder" 2013-06-19 11:08:09 +00:00
eucarc Move all EC2 cred creation to eucarc 2012-03-09 21:41:00 -06:00
exercise.sh Formatting cleanups, doc updates and whatnot 2013-04-09 14:05:32 -05:00
exerciserc make volume size parametric in boot_from_volume 2012-12-13 17:02:24 +00:00
functions Fix the OVS version check to work with upstream master versions of OVS. 2013-06-14 10:01:15 +00:00
HACKING.rst Formatting cleanups, doc updates and whatnot 2013-04-09 14:05:32 -05:00
LICENSE Add Apache 2 LICENSE file 2012-04-18 01:45:35 -05:00
openrc Functions are required to parse localrc 2013-06-01 01:36:24 -04:00
README.md Add support for setting extra networking configuration options. 2013-06-07 14:45:22 +00:00
rejoin-stack.sh Source functions in rejoin-stack.sh 2013-06-06 15:42:13 -04:00
stack.sh Install agent packages when running agents like DHCP or L3 2013-06-14 10:33:38 -07:00
stackrc Periodic formatting cleanup 2013-06-11 14:39:27 -05:00
unstack.sh On unstack : clean up network namespaces created by quantum 2013-05-22 10:00:30 +02:00

DevStack is a set of scripts and utilities to quickly deploy an OpenStack cloud.

Goals

  • To quickly build dev OpenStack environments in a clean Ubuntu or Fedora environment
  • To describe working configurations of OpenStack (which code branches work together? what do config files look like for those branches?)
  • To make it easier for developers to dive into OpenStack so that they can productively contribute without having to understand every part of the system at once
  • To make it easy to prototype cross-project features
  • To sanity-check OpenStack builds (used in gating commits to the primary repos)

Read more at http://devstack.org (built from the gh-pages branch)

IMPORTANT: Be sure to carefully read stack.sh and any other scripts you execute before you run them, as they install software and may alter your networking configuration. We strongly recommend that you run stack.sh in a clean and disposable vm when you are first getting started.

Devstack on Xenserver

If you would like to use Xenserver as the hypervisor, please refer to the instructions in ./tools/xen/README.md.

Versions

The devstack master branch generally points to trunk versions of OpenStack components. For older, stable versions, look for branches named stable/[release] in the DevStack repo. For example, you can do the following to create a diablo OpenStack cloud:

git checkout stable/diablo
./stack.sh

You can also pick specific OpenStack project releases by setting the appropriate *_BRANCH variables in localrc (look in stackrc for the default set). Usually just before a release there will be milestone-proposed branches that need to be tested::

GLANCE_REPO=https://github.com/openstack/glance.git
GLANCE_BRANCH=milestone-proposed

Start A Dev Cloud

Installing in a dedicated disposable vm is safer than installing on your dev machine! To start a dev cloud:

./stack.sh

When the script finishes executing, you should be able to access OpenStack endpoints, like so:

We also provide an environment file that you can use to interact with your cloud via CLI:

# source openrc file to load your environment with osapi and ec2 creds
. openrc
# list instances
nova list

If the EC2 API is your cup-o-tea, you can create credentials and use euca2ools:

# source eucarc to generate EC2 credentials and set up the environment
. eucarc
# list instances using ec2 api
euca-describe-instances

Customizing

You can override environment variables used in stack.sh by creating file name localrc. It is likely that you will need to do this to tweak your networking configuration should you need to access your cloud from a different host.

Database Backend

Multiple database backends are available. The available databases are defined in the lib/databases directory. mysql is the default database, choose a different one by putting the following in localrc:

disable_service mysql
enable_service postgresql

mysql is the default database.

RPC Backend

Multiple RPC backends are available. Currently, this includes RabbitMQ (default), Qpid, and ZeroMQ. Your backend of choice may be selected via the localrc.

Note that selecting more than one RPC backend will result in a failure.

Example (ZeroMQ):

ENABLED_SERVICES="$ENABLED_SERVICES,-rabbit,-qpid,zeromq"

Example (Qpid):

ENABLED_SERVICES="$ENABLED_SERVICES,-rabbit,-zeromq,qpid"

Swift

Swift is enabled by default configured with only one replica to avoid being IO/memory intensive on a small vm. When running with only one replica the account, container and object services will run directly in screen. The others services like replicator, updaters or auditor runs in background.

If you would like to disable Swift you can add this to your localrc :

disable_service s-proxy s-object s-container s-account

If you want a minimal Swift install with only Swift and Keystone you can have this instead in your localrc:

disable_all_services
enable_service key mysql s-proxy s-object s-container s-account

If you only want to do some testing of a real normal swift cluster with multiple replicas you can do so by customizing the variable SWIFT_REPLICAS in your localrc (usually to 3).

Swift S3

If you are enabling swift3 in ENABLED_SERVICES devstack will install the swift3 middleware emulation. Swift will be configured to act as a S3 endpoint for Keystone so effectively replacing the nova-objectstore.

Only Swift proxy server is launched in the screen session all other services are started in background and managed by swift-init tool.

Quantum

Basic Setup

In order to enable Quantum a single node setup, you'll need the following settings in your localrc :

disable_service n-net
enable_service q-svc
enable_service q-agt
enable_service q-dhcp
enable_service q-l3
enable_service q-meta
enable_service quantum
# Optional, to enable tempest configuration as part of devstack
enable_service tempest

Then run stack.sh as normal.

devstack supports adding specific Quantum configuration flags to both the Open vSwitch and LinuxBridge plugin configuration files. To make use of this feature, the following variables are defined and can be configured in your localrc file:

Variable Name             Plugin Config File Section Modified
-------------------------------------------------------------------------------------
Q_SRV_EXTRA_OPTS          `OVS` (for Open Vswitch) or `LINUX_BRIDGE` (for LinuxBridge)
Q_AGENT_EXTRA_AGENT_OPTS  AGENT
Q_AGENT_EXTRA_SRV_OPTS    `OVS` (for Open Vswitch) or `LINUX_BRIDGE` (for LinuxBridge)

An example of using the variables in your localrc is below:

Q_AGENT_EXTRA_AGENT_OPTS=(tunnel_type=vxlan vxlan_udp_port=8472)
Q_SRV_EXTRA_OPTS=(tenant_network_type=vxlan)

Tempest

If tempest has been successfully configured, a basic set of smoke tests can be run as follows:

$ cd /opt/stack/tempest
$ nosetests tempest/tests/network/test_network_basic_ops.py

Multi-Node Setup

A more interesting setup involves running multiple compute nodes, with Quantum networks connecting VMs on different compute nodes. You should run at least one "controller node", which should have a stackrc that includes at least:

disable_service n-net
enable_service q-svc
enable_service q-agt
enable_service q-dhcp
enable_service q-l3
enable_service q-meta
enable_service quantum

You likely want to change your localrc to run a scheduler that will balance VMs across hosts:

SCHEDULER=nova.scheduler.simple.SimpleScheduler

You can then run many compute nodes, each of which should have a stackrc which includes the following, with the IP address of the above controller node:

ENABLED_SERVICES=n-cpu,rabbit,g-api,quantum,q-agt
SERVICE_HOST=[IP of controller node]
MYSQL_HOST=$SERVICE_HOST
RABBIT_HOST=$SERVICE_HOST
Q_HOST=$SERVICE_HOST
MATCHMAKER_REDIS_HOST=$SERVICE_HOST

Cells

Cells is a new scaling option with a full spec at http://wiki.openstack.org/blueprint-nova-compute-cells.

To setup a cells environment add the following to your localrc:

enable_service n-cell
enable_service n-api-meta
MULTI_HOST=True

# The following have not been tested with cells, they may or may not work.
disable_service n-obj
disable_service cinder
disable_service c-sch
disable_service c-api
disable_service c-vol
disable_service n-xvnc

Be aware that there are some features currently missing in cells, one notable one being security groups.