Remove obsolete files from Install Guide

These files were not used at all.

Change-Id: Ic53aed04a39b14f10fb4501e61dc0b94d23f76c6
backport: stable/havana
This commit is contained in:
Andreas Jaeger 2013-10-19 13:26:35 +02:00
parent 0643f7fb25
commit d3cc4acc73
8 changed files with 0 additions and 1070 deletions

View File

@ -1,174 +0,0 @@
<?xml version="1.0" encoding="UTF-8"?>
<appendix xml:id="conf-files" xmlns="http://docbook.org/ns/docbook"
xmlns:xi="http://www.w3.org/2001/XInclude"
xmlns:xlink="http://www.w3.org/1999/xlink" version="5.0">
<title>Configuration File Examples</title>
<section xml:id="keystone-conf-file">
<title>keystone.conf</title>
<para>The configuration file for the Identity Service is
<filename class="directory"
>/etc/keystone/keystone.conf</filename>.</para>
<para>After you install the Identity Service, modify this file
to use SQL for endpoint data and to replace the ADMIN key
with the one created that was created during the
installation.</para>
<literallayout class="monospaced"><xi:include parse="text" href="../common/samples/keystone.conf"/></literallayout>
</section>
<section xml:id="keystone-paste-file">
<title>keystone-paste.ini</title>
<para>The Identity Service middleware pipeline configuration file is
<filename
class="directory">/etc/keystone/keystone-paste.ini</filename>.</para>
<para>You should not need to modify this file.</para>
<literallayout class="monospaced"><xi:include parse="text" href="samples/keystone-paste.ini"/></literallayout>
</section>
<section xml:id="glance-registry-conf-file">
<title>glance-registry.conf</title>
<para>The Image Service registry configuration file is
<filename>/etc/glance/glance-registry.conf</filename>.</para>
<para>This file stores metadata for images.</para>
<para>After you install the Image Service, modify this
configuration file.</para>
<literallayout class="monospaced"><xi:include parse="text" href="samples/glance-registry.conf"/></literallayout>
</section>
<section xml:id="glance-registry-paste-file">
<title>glance-registry-paste.ini</title>
<para>The Image Service API middleware pipeline configuration
file is <filename>/etc/glance/glance-registry-paste.ini</filename>.</para>
<para>After you install the Image Service, modify this
configuration file.</para>
<literallayout class="monospaced"><xi:include parse="text" href="samples/glance-registry-paste.ini"/></literallayout>
</section>
<section xml:id="glance-api-conf-file">
<title>glance-api.conf</title>
<para>The configuration file for the Image API is <filename>/etc/glance/glance-api.conf</filename>.</para>
<para>After you install the Image Service API, update this
file as shown in the following example.</para>
<literallayout class="monospaced"><xi:include parse="text" href="samples/glance-api.conf"/></literallayout>
</section>
<section xml:id="glance-api-paste-file">
<title>glance-api-paste.ini</title>
<para>The Image Service API middleware pipeline configuration
file is <filename>/etc/glance/glance-api-paste.ini</filename>.</para>
<para>You should not need to modify this file.</para>
<literallayout class="monospaced"><xi:include parse="text" href="samples/glance-api-paste.ini"/></literallayout>
</section>
<section xml:id="glance-scrubber-conf-file">
<title>glance-scrubber.conf</title>
<para>An additional configuration file for the Image Service
is <filename class="directory"
>/etc/glance/glance-scrubber.conf</filename>.</para>
<para>The scrubber is a utility that cleans up images that
have been deleted.</para>
<literallayout class="monospaced"><xi:include parse="text" href="samples/glance-scrubber.conf"/></literallayout>
</section>
<section xml:id="nova-conf-file">
<title>nova.conf</title>
<para>The Compute configuration file is <filename
class="directory">/etc/nova/nova.conf</filename>.</para>
<para>For a list of configuration options for this file, see
the List of Tables on this page: <link
xlink:href="http://docs.openstack.org/trunk/config-reference/content/"
><citetitle>OpenStack Configuration Reference</citetitle></link>.</para>
<para>This guide assumes that the IP address of the machine
that runs the Identity Service is
<literal>192.168.206.130</literal>. If the IP address
of the machine on your setup is different, change the
following configuration options:<itemizedlist>
<listitem>
<para><literal>service_host</literal></para>
</listitem>
<listitem>
<para><literal>auth_host</literal></para>
</listitem>
<listitem>
<para><literal>auth_uri</literal></para>
</listitem>
</itemizedlist></para>
<literallayout os="ubuntu;opensuse;sles" class="monospaced"><xi:include parse="text" href="../common/samples/nova.conf"/></literallayout>
<literallayout os="rhel;centos;fedora" class="monospaced"><xi:include parse="text" href="../common/samples/nova.conf-yum"/></literallayout>
</section>
<section xml:id="api-paste-ini-file">
<title>api-paste.ini</title>
<para>The middleware configuration file used by the EC2 API
and OpenStack Compute API is <filename class="directory"
>/etc/nova/api-paste.ini</filename>.</para>
<para>You should not need to edit it.</para>
<literallayout class="monospaced"> <xi:include parse="text" href="samples/api-paste.ini"/></literallayout>
</section>
<section xml:id="novarc-file">
<title>Credentials (openrc)</title>
<para>This file contains the credentials used by Compute,
Image, and Identity services. You can optionally store
this file in<filename class="directory">
/home/openrc</filename>.</para>
<para>Do not source this file in the environment from where
you issue commands.</para>
<para>Run "env | grep OS_" or "env | grep NOVA_" to view what
is being used in your environment.</para>
<literallayout class="monospaced"><xi:include parse="text" href="samples/openrc.txt"/></literallayout>
</section>
<section xml:id="cinder-conf">
<title>cinder.conf</title>
<literallayout><xi:include parse="text" href="../common/samples/cinder.conf"/></literallayout>
</section>
<section xml:id="local-settings-py-file">
<title>Dashboard configuration</title>
<para>This file contains the database and configuration
settings for the OpenStack Dashboard.</para>
<literallayout class="monospaced"><xi:include parse="text" href="../common/samples/local_settings.py"/></literallayout>
</section>
<section xml:id="swift-conf">
<title>etc/swift/swift.conf</title>
<para>This file contains the settings to randomize the hash
for the ring for Object Storage, code-named swift.</para>
<literallayout class="monospaced"><xi:include parse="text" href="samples/swift.conf.txt"/></literallayout>
</section>
<section xml:id="network-interfaces-conf">
<title>etc/network/interfaces.conf</title>
<para>These instructions are for using the FlatDHCP networking
mode with a single network interface.</para>
<literallayout class="monospaced"><xi:include parse="text" href="samples/network-interfaces.conf.txt"/></literallayout>
</section>
<section xml:id="swift-proxy-conf">
<title>etc/swift/proxy-server.conf</title>
<para>This file contains the settings for the Object Storage
proxy server, which contains the Identity service
settings.</para>
<literallayout class="monospaced"><xi:include parse="text" href="samples/proxy-server.conf.txt"/></literallayout>
</section>
<section xml:id="account-server-conf">
<title>etc/swift/account-server.conf</title>
<literallayout class="monospaced"><xi:include parse="text" href="samples/account-server.conf.txt"/></literallayout>
</section>
<section xml:id="account-server-conf-1">
<title>etc/swift/account-server/1.conf</title>
<literallayout class="monospaced"><xi:include parse="text" href="samples/account-server-1.conf.txt"/></literallayout>
</section>
<section xml:id="container-server-conf">
<title>etc/swift/container-server.conf</title>
<literallayout class="monospaced"><xi:include parse="text" href="samples/container-server.conf.txt"/></literallayout>
</section>
<section xml:id="container-server-conf-1">
<title>etc/swift/container-server/1.conf</title>
<literallayout class="monospaced"><xi:include parse="text" href="samples/container-server-1.conf.txt"/></literallayout>
</section>
<section xml:id="object-server-conf">
<title>etc/swift/object-server.conf</title>
<literallayout class="monospaced"><xi:include parse="text" href="samples/object-server.conf.txt"/></literallayout>
</section>
<section xml:id="object-server-conf-1">
<title>etc/swift/object-server/1.conf</title>
<literallayout class="monospaced"><xi:include parse="text" href="samples/object-server-1.conf.txt"/></literallayout>
</section>
</appendix>

View File

@ -1,13 +0,0 @@
<?xml version="1.0" encoding="UTF-8"?>
<appendix xml:id="dependencies"
xmlns="http://docbook.org/ns/docbook"
xmlns:xi="http://www.w3.org/2001/XInclude"
xmlns:xlink="http://www.w3.org/1999/xlink" version="5.0">
<title>Appendix: Dependencies</title>
<para>Included for your reference is a list of dependencies.</para>
<section xml:id="ubuntu-dependencies"><title>Ubuntu 12.04 dependencies</title><para></para>
<literallayout class="monospaced"><xi:include parse="text"
href="samples/ubuntu_dependencies.txt"/></literallayout></section>
</appendix>

View File

@ -1,94 +0,0 @@
<?xml version="1.0" encoding="UTF-8"?>
<chapter xmlns="http://docbook.org/ns/docbook"
xmlns:xi="http://www.w3.org/2001/XInclude"
xmlns:xlink="http://www.w3.org/1999/xlink" version="5.0"
xml:id="basic-install_architecture">
<title>Architecture</title>
<figure><title>Physical network diagram</title> <mediaobject>
<imageobject>
<imagedata width="6in" fileref="../../common/figures/Neutron-PhysNet-Diagram.png"/>
</imageobject>
<!-- Edit link: http://docs.google.com/drawings/d/19qtlB_EeG9TM_jpeX5DprHRd9ShngXgfpt_t5mlrUMo/edit -->
</mediaobject></figure>
<para>This guide creates the following types of OpenStack
servers:</para>
<itemizedlist>
<listitem>
<para><emphasis role="bold">The Cloud
Controller</emphasis>. Provides all functionality
of the cloud except actually hosting virtual machines
or providing network services. See the "Compute Node"
and "Network Controller" for details about those
roles. This server hosts the OpenStack Image
Service, Block Storage Service, Identity Service, and the
dashboard. It also runs portions of the OpenStack
Compute service, such as the API server, the scheduler,
conductor, console authenticator, and VNC service.
Finally, it hosts the API endpoint for the OpenStack
Network service.</para>
</listitem>
<listitem>
<para><emphasis role="bold">The Network
Controller</emphasis>. Provides the bulk of the
OpenStack Network services such as DHCP, layer 2
switching, layer 3 routing, floating IPs (which this
guide does not configure), and metadata connectivity.</para>
</listitem>
<listitem>
<para><emphasis role="bold">Compute Node</emphasis>. Runs
the OpenStack Compute service as well as the OpenStack
Network service agent (in this case, the Open vSwitch
plugin agent). This server also manages an
OpenStack-compatible hypervisor such as KVM or Xen.
This server hosts the actual virtual machines
(instances).</para>
</listitem>
</itemizedlist>
<note><para>OpenStack provides great flexibility with regard to how its
individual services can be hosted. For example, the services
that run on the Network Controller can easily be installed on
the Cloud Controller. As another example, the OpenStack Image
service can be installed on its own server (or many servers to
provide a more highly available service).</para>
</note>
<para>With regard to cloud networking, a standard OpenStack
Network setup can have up to four distinct physical data
center networks. Note that these networks can be combined and
re-used. For example, the Management, Data, and API networks
are commonly the same network. For simplicity, this guide shows this configuration.</para>
<itemizedlist>
<listitem>
<para><emphasis role="bold">Management
network</emphasis>. Used for internal
communication between OpenStack components. The IP
addresses on this network should be reachable only
within the data center.</para>
</listitem>
<listitem>
<para><emphasis role="bold">Data
network</emphasis>. Used for VM data communication
within the cloud deployment. The IP addressing
requirements of this network depend on the OpenStack
Networking plugin in use.</para>
</listitem>
<listitem>
<para><emphasis role="bold">External
network</emphasis>. Provides VMs with Internet
access in some deployment scenarios. The IP addresses
on this network should be reachable by anyone on the
Internet.</para>
</listitem>
<listitem>
<para><emphasis role="bold">API
network</emphasis>. Exposes all OpenStack APIs,
including the OpenStack Networking API, to
tenants. The IP addresses on this network should be
reachable by anyone on the Internet. This may be the
same network as the external network, as it is
possible to create a subnet for the external
network that uses IP allocation ranges to use only
less than the full range of IP addresses in an IP
block.</para>
</listitem>
</itemizedlist>
</chapter>

View File

@ -1,99 +0,0 @@
<?xml version="1.0" encoding="UTF-8"?>
<section xmlns="http://docbook.org/ns/docbook"
xmlns:xi="http://www.w3.org/2001/XInclude"
xmlns:xlink="http://www.w3.org/1999/xlink" version="5.0"
xml:id="basic-install_compute-neutron">
<title>OpenStack Networking (Compute Node)</title>
<section xml:id="compute-ovs">
<title>Open vSwitch</title>
<para>
<orderedlist>
<listitem>
<para>Install the packages:
<screen os="ubuntu;debian"><prompt>#</prompt> <userinput>apt-get install openvswitch-switch</userinput></screen>
<screen os="centos;rhel;fedora"><prompt>#</prompt> <userinput>yum install openvswitch-switch</userinput></screen>
<screen os="opensuse"><prompt>#</prompt> <userinput>zypper install openvswitch-switch</userinput></screen>
</para>
</listitem>
<listitem>
<para>Start Open vSwitch service
<screen os="ubuntu;debian"><prompt>#</prompt> <userinput>service openvswitch-switch start</userinput></screen>
<screen os="centos;rhel;fedora"><prompt>#</prompt> <userinput>service openvswitch-switch start</userinput>
<prompt>#</prompt> <userinput>chkconfig openvswitch-switch on</userinput></screen>
<screen os="opensuse"><prompt>#</prompt> <userinput>systemctl start openvswitch-switch</userinput>
<prompt>#</prompt> <userinput>systemctl enable openvswitch-switch</userinput></screen>
</para>
</listitem>
<listitem>
<para>Create an internal bridge. Just as described in the Introduction to this guide, the Compute Node
does not provide an external bridge. This enforces all instances' network traffic to go through
the Network Controller. This is known as a "single-node" networking setup.
<screen><prompt>#</prompt> <userinput>ovs-vsctl add-br br-int</userinput></screen>
</para>
</listitem>
</orderedlist>
</para>
</section>
<section xml:id="compute-neutron">
<title>OpenStack Networking</title>
<para>
<orderedlist>
<listitem>
<para>Install the packages:
<screen os="ubuntu;debian"><prompt>#</prompt> <userinput>apt-get install neutron-plugin-openvswitch-agent</userinput></screen>
<screen os="centos;rhel;fedora"><prompt>#</prompt> <userinput>yum install openstack-neutron-openvswitch</userinput></screen>
<screen os="opensuse"><prompt>#</prompt> <userinput>zypper install openstack-neutron-openvswitch-agent</userinput></screen>
</para>
</listitem>
<listitem>
<para>Edit <filename>/etc/neutron/neutron.conf</filename>:
<programlisting os="ubuntu;debian" language="ini">[DEFAULT]
rabbit_host = 10.10.10.10
rabbit_password = password
verbose = True</programlisting>
<programlisting os="centos;rhel;fedora" language="ini">verbose = True
rpc_backend=neutron.openstack.common.rpc.impl_qpid</programlisting>
<programlisting os="opensuse" language="ini">[DEFAULT]
rabbit_host = 10.10.10.10
rabbit_password = password
verbose = True</programlisting>
</para>
</listitem>
<listitem>
<para>Edit <filename>/etc/neutron/plugins/openvswitch/ovs_neutron_plugin.ini</filename>:
<programlisting language="ini">[database]
connection = mysql://neutron:password@10.10.10.1/neutron
[ovs]
tenant_network_type = gre
tunnel_id_ranges = 1:1000
local_ip = 10.10.10.11
enable_tunneling = True
[securitygroup]
firewall_driver = \
neutron.agent.linux.iptables_firewall.OVSHybridIptablesFirewallDriver</programlisting>
<screen os="centos;rhel;fedora"><prompt>#</prompt> <userinput>ln -s /etc/neutron/plugins/openvswitch/ovs_neutron_plugin.ini \
/etc/neutron/plugin.ini</userinput></screen>
<screen os="opensuse"><prompt>#</prompt> <userinput>ln -s /etc/neutron/plugins/openvswitch/ovs_neutron_plugin.ini \
/etc/neutron/plugin.ini</userinput></screen>
</para>
</listitem>
<listitem>
<para>Start the Agent:
<screen os="ubuntu;debian"><prompt>#</prompt> <userinput>service neutron-plugin-openvswitch-agent restart</userinput></screen>
<screen os="centos;rhel;fedora"><prompt>#</prompt> <userinput>service neutron-openvswitch-agent restart</userinput>
<prompt>#</prompt> <userinput>chkconfig neutron-openvswitch-agent on</userinput></screen>
<screen os="opensuse"><prompt>#</prompt> <userinput>systemctl restart openstack-neutron-openvswitch-agent</userinput>
<prompt>#</prompt> <userinput>systemctl enable openstack-neutron-openvswitch-agent</userinput></screen>
</para>
</listitem>
<listitem os="centos;rhel;fedora">
<para>Ensure the cleanup utility is started on future boots:</para>
<screen><prompt>#</prompt> <userinput>chkconfig neutron-ovs-cleanup on</userinput></screen>
</listitem>
</orderedlist>
<note><para>Check the <filename>/var/log/neutron/openvswitch-agent.log</filename> file for errors that would
prevent the Networking service from successfully starting.</para></note>
</para>
</section>
</section>

View File

@ -1,133 +0,0 @@
<?xml version="1.0" encoding="UTF-8"?>
<section xmlns="http://docbook.org/ns/docbook"
xmlns:xi="http://www.w3.org/2001/XInclude"
xmlns:xlink="http://www.w3.org/1999/xlink" version="5.0"
xml:id="basic-install_controller-cinder">
<title>OpenStack Block Storage</title>
<para>While OpenStack Block Storage contains many different storage drivers, the
most basic configuration uses LVM and iSCSI. This
guide illustrates how to use one disk
(<literal>/dev/sdb</literal>) in an LVM Volume Group
called <literal>cinder-volumes</literal>. When a user requests
a block storage volume, a Logical Volume is created from this
Volume Group and then mounted on the user's instance by way of
iSCSI. <orderedlist>
<listitem>
<para>Install the OpenStack Block Storage packages:
<screen os="ubuntu;debian"><prompt>#</prompt> <userinput>apt-get install cinder-api cinder-scheduler cinder-volume iscsitarget \
open-iscsi iscsitarget-dkms python-cinderclient \
linux-headers-`uname -r`</userinput></screen>
<screen os="centos;rhel;fedora"><prompt>#</prompt> <userinput>yum install openstack-cinder openstack-cinder-doc \
iscsi-initiator-utils scsi-target-utils</userinput></screen>
<screen os="opensuse"><prompt>#</prompt> <userinput>zypper install openstack-cinder-api openstack-cinder-scheduler \
openstack-cinder-volume open-iscsi iscsitarget</userinput></screen>
</para>
</listitem>
<listitem>
<para>Configure &amp; start the iSCSI services: <screen os="ubuntu;debian"><prompt>#</prompt> <userinput>sed -i 's/false/true/g' /etc/default/iscsitarget</userinput>
<prompt>#</prompt> <userinput>service iscsitarget start</userinput>
<prompt>#</prompt> <userinput>service open-iscsi start</userinput></screen>
<screen os="centos;rhel;fedora"><prompt>#</prompt> <userinput>service tgtd start</userinput>
<prompt>#</prompt> <userinput>service iscsi start</userinput>
<prompt>#</prompt> <userinput>chkconfig tgtd on</userinput>
<prompt>#</prompt> <userinput>chkconfig iscsi on</userinput></screen>
<screen os="opensuse"><prompt>#</prompt> <userinput>systemctl start iscsitarget.service iscsid.service</userinput>
<prompt>#</prompt> <userinput>systemctl enable iscsitarget.service iscsid.service</userinput></screen>
</para>
</listitem>
<listitem>
<para>Configure OpenStack Block Storage:</para>
<itemizedlist>
<listitem>
<para>Edit <filename>/etc/cinder/cinder.conf</filename>:
<programlisting os="ubuntu;debian" language="ini">[DEFAULT]
sql_connection = mysql://cinder:password@localhost/cinder
rabbit_password = password
[keystone_authtoken]
admin_tenant_name = service
admin_user = cinder
admin_password = password
service_protocol = http
service_host = localhost
service_port = 5000
auth_host = localhost
auth_port = 35357
auth_protocol = http</programlisting><programlisting os="centos;rhel;fedora" language="ini">[DEFAULT]
rpc_backend = cinder.openstack.common.rpc.impl_qpid
sql_connection = mysql://cinder:password@localhost/cinder
qpid_user = guest
qpid_password = quest
[keystone_authtoken]
admin_tenant_name = service
admin_user = cinder
admin_password = password
service_protocol = http
service_host = localhost
service_port = 5000
auth_host = localhost
auth_port = 35357
auth_protocol = http</programlisting>
<programlisting os="opensuse" language="ini">[DEFAULT]
sql_connection = mysql://cinder:password@localhost/cinder
rabbit_password = password
[keystone_authtoken]
admin_tenant_name = service
admin_user = cinder
admin_password = password
service_protocol = http
service_host = localhost
service_port = 5000
auth_host = localhost
auth_port = 35357
auth_protocol = http</programlisting>
</para>
</listitem>
<listitem>
<para>Edit <filename>/etc/cinder/api-paste.ini</filename> in the
<literal>filter_authtoken</literal> section so it does not contain
any of the variables defined in the
<literal>keystone_authtoken</literal> section of
<filename>cinder.conf</filename>:
<programlisting language="ini">[filter:authtoken]
paste.filter_factory = keystoneclient.middleware.auth_token:filter_factory
# This section should not contain any other config options</programlisting>
</para>
</listitem>
<listitem>
<para>Create the LVM Physical Volume and
Logical Volume:
<screen><prompt>#</prompt> <userinput>pvcreate /dev/sdb</userinput>
<prompt>#</prompt> <userinput>vgcreate cinder-volumes /dev/sdb</userinput></screen>
</para>
</listitem>
<listitem>
<para>Create the OpenStack Block Storage tables in the database:
<screen><prompt>#</prompt> <userinput>cinder-manage db sync</userinput></screen>
</para>
</listitem>
<listitem>
<para>Restart the services: <screen os="ubuntu;debian"><prompt>#</prompt> <userinput>service cinder-api restart</userinput>
<prompt>#</prompt> <userinput>service cinder-scheduler restart</userinput>
<prompt>#</prompt> <userinput>service cinder-volume restart</userinput></screen>
<screen os="centos;rhel;fedora"><prompt>#</prompt> <userinput>service openstack-cinder-api restart</userinput>
<prompt>#</prompt> <userinput>service openstack-cinder-scheduler restart</userinput>
<prompt>#</prompt> <userinput>service openstack-cinder-volume restart</userinput>
<prompt>#</prompt> <userinput>chkconfig openstack-cinder-api on</userinput>
<prompt>#</prompt> <userinput>chkconfig openstack-cinder-scheduler on</userinput>
<prompt>#</prompt> <userinput>chkconfig openstack-cinder-volume on</userinput></screen>
<screen os="opensuse"><prompt>#</prompt> <userinput>systemctl restart openstack-cinder-api</userinput>
<prompt>#</prompt> <userinput>systemctl restart openstack-cinder-scheduler.service</userinput>
<prompt>#</prompt> <userinput>systemctl restart openstack-cinder-volume.service</userinput>
<prompt>#</prompt> <userinput>systemctl enable openstack-cinder-api.service</userinput>
<prompt>#</prompt> <userinput>systemctl enable openstack-cinder-scheduler.service</userinput>
<prompt>#</prompt> <userinput>systemctl enable openstack-cinder-volume.service</userinput></screen></para>
</listitem>
</itemizedlist>
</listitem>
</orderedlist>
</para>
</section>

View File

@ -1,98 +0,0 @@
<?xml version="1.0" encoding="UTF-8"?>
<section xmlns="http://docbook.org/ns/docbook"
xmlns:xi="http://www.w3.org/2001/XInclude"
xmlns:xlink="http://www.w3.org/1999/xlink" version="5.0"
xml:id="basic-install_network-operating">
<title>Virtual Networking</title>
<section xml:id="create-networking">
<title>Create Virtual Networking</title>
<para>
<orderedlist>
<listitem>
<para>Create an <literal>openrc</literal> File</para>
<para>
<itemizedlist>
<listitem>
<para>Create a file called
<filename>~/openrc</filename>. This
file contains the OpenStack admin
credentials that are used when
interacting with the OpenStack
environment on the command line.
<programlisting language="bash">export OS_TENANT_NAME=admin
export OS_USERNAME=admin
export OS_PASSWORD=password
export OS_AUTH_URL="http://10.10.10.10:5000/v2.0/"
export SERVICE_ENDPOINT="http://10.10.10.10:35357/v2.0"
export SERVICE_TOKEN=password</programlisting></para>
</listitem>
</itemizedlist>
</para>
<itemizedlist>
<listitem>
<para>Source the credentials into your environment:
<screen><prompt>$</prompt> <userinput>source ~/openrc</userinput></screen></para></listitem>
<listitem>
<para>
Configure the Bash shell to load these credentials upon each login:
<screen><prompt>$</prompt> <userinput>echo "source ~/openrc" >> ~/.bashrc</userinput></screen>
</para>
</listitem>
</itemizedlist>
</listitem>
<listitem>
<para>The following bash script creates an internal network for the "demo" project.
<programlisting language="bash">#!/bin/bash
TENANT_NAME="demo"
TENANT_NETWORK_NAME="demo-net"
TENANT_SUBNET_NAME="${TENANT_NETWORK_NAME}-subnet"
TENANT_ROUTER_NAME="demo-router"
FIXED_RANGE="10.5.5.0/24"
NETWORK_GATEWAY="10.5.5.1"
TENANT_ID=$(keystone tenant-list | grep " $TENANT_NAME " | awk '{print $2}')
TENANT_NET_ID=$(neutron net-create --tenant_id $TENANT_ID \
$TENANT_NETWORK_NAME --provider:network_type gre \
--provider:segmentation_id 1 | grep " id " | awk '{print $4}')
TENANT_SUBNET_ID=$(neutron subnet-create --tenant_id $TENANT_ID \
--ip_version 4 --name $TENANT_SUBNET_NAME $TENANT_NET_ID $FIXED_RANGE \
--gateway $NETWORK_GATEWAY --dns_nameservers list=true 8.8.8.8 | \
grep " id " | awk '{print $4}')
ROUTER_ID=$(neutron router-create --tenant_id $TENANT_ID \
$TENANT_ROUTER_NAME | grep " id " | awk '{print $4}')
neutron router-interface-add $ROUTER_ID $TENANT_SUBNET_ID</programlisting>
</para>
</listitem>
</orderedlist>
</para>
</section>
<section xml:id="configure-l3">
<title>L3 Configuration</title>
<para>The OpenStack Networking L3 service enables instances to have external network access. If this service is not configured, your instances
can only communicate with each other. Note that this configuration is highly dependant on your environment.
For example, make note of the <literal>subnet-create</literal> command below. You must verify your own network settings
for the external subnet (<literal>10.0.0.0/24</literal> in this case) as well as an allocation pool. The allocation pool
is used to provide each Project with an IP address to access the external network. The pool consists of 50 IPs and therefore
only 50 projects can get a gateway IP.</para>
<itemizedlist>
<listitem>
<para>Create an external network:
<screen><prompt>$</prompt> <userinput>neutron net-create public --router:external=True</userinput></screen>
</para>
</listitem>
<listitem>
<para>Create a subnet for the external network:
<screen><prompt>$</prompt> <userinput>neutron subnet-create --ip_version 4 \
--gateway 10.0.0.1 public 10.0.0.0/24 \
--allocation-pool start=10.0.0.200,end=10.0.0.250 --disable-dhcp \
--name public-subnet</userinput></screen>
</para>
</listitem>
<listitem>
<para>Set the gateway of the demo router to the public network:
<screen><prompt>$</prompt> <userinput>neutron router-gateway-set demo-router public</userinput></screen>
</para>
</listitem>
</itemizedlist>
</section>
</section>

View File

@ -1,224 +0,0 @@
<?xml version="1.0" encoding="UTF-8"?>
<section xmlns="http://docbook.org/ns/docbook"
xmlns:xi="http://www.w3.org/2001/XInclude"
xmlns:xlink="http://www.w3.org/1999/xlink" version="5.0"
xml:id="basic-install_network-services">
<title>OpenStack Networking (Network Controller)</title>
<section xml:id="network-ovs">
<title>Open vSwitch</title>
<procedure>
<title>To configure Open vSwitch</title><step>
<para>Install the packages:</para>
<screen os="ubuntu;debian"><prompt>#</prompt> <userinput>apt-get install neutron-plugin-openvswitch-agent \
neutron-dhcp-agent neutron-l3-agent</userinput></screen>
<screen os="centos;rhel;fedora"><prompt>#</prompt> <userinput>yum install openstack-neutron openstack-neutron-openvswitch\
openvswitch-switch</userinput></screen>
<screen os="opensuse"><prompt>#</prompt> <userinput>zypper install openstack-neutron openstack-neutron-openvswitch-agent \
openvswitch-switch</userinput></screen>
<para os="centos;rhel;fedora">In Fedora, the Open
vSwitch kernel module is compiled without tunnel
support. If you need gre tunnels and network
namespaces, you must recompile this package from
source. For information, see the Open vSwitch
site. If you recompile the kernel module, you must
remove and delete the existing one.</para>
</step>
<step>
<para>Start Open vSwitch:</para>
<screen os="centos;rhel;fedora"><prompt>#</prompt> <userinput>service openvswitch-switch start</userinput>
<prompt>#</prompt> <userinput>chkconfig openvswitch-switch on</userinput></screen>
<screen os="ubuntu;debian"><prompt>#</prompt> <userinput>service openvswitch-switch start</userinput></screen>
<screen os="opensuse"><prompt>#</prompt> <userinput>systemctl start openvswitch-switch.service</userinput>
<prompt>#</prompt> <userinput>systemctl enable openvswitch-switch.service</userinput></screen>
</step>
<step>
<para>Create an internal and external network bridge.
</para><note><para>The introduction in this guide describes the
purposes of these bridges.</para></note>
<screen><prompt>#</prompt> <userinput>ovs-vsctl add-br br-ex</userinput>
<prompt>#</prompt> <userinput>ovs-vsctl add-port br-ex eth1</userinput>
<prompt>#</prompt> <userinput>ovs-vsctl add-br br-int</userinput></screen>
</step>
<step>
<para>To configure the bridges, complete the
following steps:</para>
<substeps>
<step os="ubuntu;debian">
<para>Change the <literal>eth1</literal>
entry in <filename>
/etc/network/interfaces</filename>, as
follows:</para>
<programlisting>auto eth1
iface eth1 inet manual
up ip address add 0/0 dev $IFACE
up ip link set $IFACE up
down ip link set $IFACE down</programlisting>
</step>
<step os="ubuntu;debian">
<para>Add <literal>br-ex</literal> to
<filename>/etc/network/interfaces</filename>,
as follows:</para>
<programlisting>auto br-ex
iface br-ex inet static
address 10.0.0.9
netmask 255.255.255.0
gateway 10.0.0.1</programlisting>
</step>
<step os="centos;rhel;fedora">
<para>Edit
<filename>/etc/sysconf/network-scripts/ifcfg-eth1</filename>,
as follows:</para>
<programlisting># External
DEVICE=eth1
TYPE=Ethernet
BOOTPROTO=none
NM_CONTROLLED=no
BRIDGE=br-ex
ONBOOT=yes</programlisting>
</step>
<step os="centos;rhel;fedora">
<para>Edit
<filename>/etc/sysconf/network-scripts/ifcfg-br-ex</filename>,
as follows:</para>
<programlisting>#Public Bridge
DEVICE=br-ex
TYPE=Bridge
BOOTPROTO=static
IPADDR=10.10.10.9
NETMASK=255.255.255.0
NM_CONTROLLED=no
ONBOOT=yes</programlisting>
</step>
<step>
<para>Remove the IP address from
<literal>eth1</literal> add it to
<literal>br-ex</literal>, as
follows:</para>
<screen><prompt>#</prompt> <userinput>ip addr del 10.0.0.9/24 dev eth1</userinput>
<prompt>#</prompt> <userinput>ip addr add 10.0.0.9/24 dev br-ex</userinput></screen></step>
<step os="ubuntu;debian">
<para>Restart networking, as
follows:</para>
<screen><prompt>#</prompt> <userinput>service networking restart</userinput></screen>
</step>
</substeps>
</step>
<step>
<para>Enable a simple NAT service so that
Compute nodes can access the Internet through the
Cloud Controller:</para>
<screen><prompt>#</prompt> <userinput>iptables -A FORWARD -i eth0 -o br-ex -s 10.10.10.0/24 -m conntrack \
--ctstate NEW -j ACCEPT</userinput>
<prompt>#</prompt> <userinput>iptables -A FORWARD -m conntrack --ctstate ESTABLISHED,RELATED -j ACCEPT</userinput>
<prompt>#</prompt> <userinput>iptables -A POSTROUTING -s 10.10.10.0/24 -t nat -j MASQUERADE</userinput></screen>
<note><para>These rules are lost on reboot. To save or restore the rules, use an appropriate tool from your
distribution to save and restore them.</para></note>
</step>
</procedure>
</section>
<section xml:id="network-neutron">
<title>OpenStack Networking</title>
<procedure><title>To configure the OpenStack Networking services</title>
<step>
<para>Edit
<filename>/etc/neutron/neutron.conf</filename>,
as follows:
<programlisting os="ubuntu;debian" language="ini">[DEFAULT]
verbose = True
rabbit_password = password
rabbit_host = 10.10.10.10
[keystone_authtoken]
auth_host = 10.10.10.10
admin_tenant_name = service
admin_user = neutron
admin_password = password</programlisting><programlisting os="centos;rhel;fedora" language="ini">[DEFAULT]
verbose = True
rpc_backend = cinder.openstack.common.rpc.impl_qpid
sql_connection = mysql://cinder:password@localhost/cinder
qpid_user = guest
qpid_password = quest
[keystone_authtoken]
auth_host = 10.10.10.10
admin_tenant_name = service
admin_user = neutron
admin_password = password</programlisting><programlisting os="opensuse" language="ini">[DEFAULT]
verbose = True
rabbit_password = password
rabbit_host = 10.10.10.10
[keystone_authtoken]
auth_host = 10.10.10.10
admin_tenant_name = service
admin_user = neutron
admin_password = password</programlisting></para>
</step>
<step>
<para>Edit
<filename>/etc/neutron/plugins/openvswitch/ovs_neutron_plugin.ini</filename>,
as follows: <programlisting language="ini">[database]
connection = mysql://neutron:password@10.10.10.10/neutron
[ovs]
tenant_network_type = gre
tunnel_id_ranges = 1:1000
enable_tunneling = True
local_ip = 10.10.10.9
[securitygroup]
firewall_driver = \
neutron.agent.linux.iptables_firewall.OVSHybridIptablesFirewallDriver</programlisting>
<note xmlns:db="http://docbook.org/ns/docbook">
<para>Choose <emphasis role="bold">tunnel
mode</emphasis> because you do not
have to configure your physical switches
for VLANs.</para>
</note></para>
</step>
<step>
<para>Edit
<filename>/etc/neutron/dhcp_agent.ini</filename>,
as follows:
<programlisting language="ini">[DEFAULT]
enable_isolated_metadata = True
enable_metadata_network = True</programlisting></para>
</step>
<step>
<para>Edit
<filename>/etc/neutron/metadata_agent.ini</filename>,
as follows:
<programlisting language="ini">[DEFAULT]
auth_url = http://10.10.10.10:35357/v2.0
auth_region = RegionOne
admin_tenant_name = service
admin_user = neutron
admin_password = password
nova_metadata_ip = 10.10.10.10
metadata_proxy_shared_secret = password</programlisting></para>
</step>
<step><para>Start the services, as follows:</para>
<screen os="ubuntu;debian"><prompt>#</prompt> <userinput>service neutron-plugin-openvswitch-agent start</userinput>
<prompt>#</prompt> <userinput>service neutron-dhcp-agent restart</userinput>
<prompt>#</prompt> <userinput>service neutron-metadata-agent restart</userinput>
<prompt>#</prompt> <userinput>service neutron-l3-agent restart</userinput></screen>
<screen os="centos;rhel;fedora"><prompt>#</prompt> <userinput>service neutron-server restart</userinput>
<prompt>#</prompt> <userinput>service neutron-openvswitch-agent restart</userinput>
<prompt>#</prompt> <userinput>service neutron-dhcp-agent restart</userinput>
<prompt>#</prompt> <userinput>service neutron-l3-agent restart</userinput>
<prompt>#</prompt> <userinput>chkconfig neutron-server on</userinput>
<prompt>#</prompt> <userinput>chkconfig neutron-openvswitch-agent on</userinput>
<prompt>#</prompt> <userinput>chkconfig neutron-dhcp-agent on</userinput>
<prompt>#</prompt> <userinput>chkconfig neutron-l3-agent on</userinput></screen>
<screen os="opensuse"><prompt>#</prompt> <userinput>systemctl restart neutron-server</userinput>
<prompt>#</prompt> <userinput>systemctl restart neutron-openvswitch-agent</userinput>
<prompt>#</prompt> <userinput>systemctl restart neutron-dhcp-agent</userinput>
<prompt>#</prompt> <userinput>systemctl restart neutron-l3-agent</userinput>
<prompt>#</prompt> <userinput>systemctl enable neutron-server</userinput>
<prompt>#</prompt> <userinput>systemctl enable neutron-openvswitch-agent</userinput>
<prompt>#</prompt> <userinput>systemctl enable neutron-dhcp-agent</userinput>
<prompt>#</prompt> <userinput>systemctl enable neutron-l3-agent</userinput></screen>
<note>
<para>To debug errors that prevent the Networking Service from
starting, review the
<filename>/var/log/neutron/*.log</filename>
files.</para>
</note></step>
</procedure>
</section>
</section>

View File

@ -1,235 +0,0 @@
os|distro=precise
os|vendor=Ubuntu
os|release=12.04
git|cinder|master[45e67e8]
git|glance|master[923b901]
git|horizon|master[b8c2634]
git|keystone|master[f261f71]
git|nova|master[737ca6e]
git|noVNC|master[d5fe150]
git|python-cinderclient|master[45bcdca]
git|python-glanceclient|master[e140dbb]
git|python-keystoneclient|master[0a8c960]
git|python-novaclient|master[bfb0f70]
git|python-openstackclient|master[cd14009]
pkg|bridge-utils|1.5-2ubuntu6
pkg|pep8|0.6.1-2ubuntu2
pkg|pylint|0.25.0-1ubuntu1
pkg|python-pip|1.0-1build1
pkg|screen|4.0.3-14ubuntu8
pkg|unzip|6.0-4ubuntu1
pkg|wget|1.13.4-2ubuntu1
pkg|psmisc|22.15-2ubuntu1.1
pkg|git-core|1:1.7.9.5-1
pkg|lsof|4.81.dfsg.1-1build1
pkg|openssh-server|1:5.9p1-5ubuntu1
pkg|vim-nox|2:7.3.429-2ubuntu2.1
pkg|locate|4.4.2-4ubuntu1
pkg|python-virtualenv|1.7.1.2-1
pkg|python-unittest2|0.5.1-1ubuntu1
pkg|iputils-ping|3:20101006-1ubuntu1
pkg|wget|1.13.4-2ubuntu1
pkg|curl|7.22.0-3ubuntu4
pkg|tcpdump|4.2.1-1ubuntu2
pkg|euca2ools|2.0.0~bzr516-0ubuntu3.1
pkg|tar|1.26-4ubuntu1
pkg|python-cmd2|0.6.3-1
pkg|gcc|4:4.6.3-1ubuntu5
pkg|libxml2-dev|2.7.8.dfsg-5.1ubuntu4.1
pkg|python-dev|2.7.3-0ubuntu2
pkg|python-eventlet|0.9.16-1ubuntu4.1
pkg|python-routes|1.12.3-1ubuntu1
pkg|python-greenlet|0.3.1-1ubuntu5.1
pkg|python-sqlalchemy|0.7.4-1ubuntu0.1
pkg|python-wsgiref|
pkg|python-pastedeploy|1.5.0-2build1
pkg|python-xattr|0.6.2-1ubuntu1
pkg|python-iso8601|0.1.4-1ubuntu1
pkg|python-setuptools|0.6.24-1ubuntu1
pkg|python-dev|2.7.3-0ubuntu2
pkg|python-lxml|2.3.2-1
pkg|python-pastescript|1.7.5-1
pkg|python-pastedeploy|1.5.0-2build1
pkg|python-paste|1.7.5.1-4ubuntu2
pkg|sqlite3|3.7.9-2ubuntu1.1
pkg|python-pysqlite2|2.6.3-2build1
pkg|python-sqlalchemy|0.7.4-1ubuntu0.1
pkg|python-mysqldb|1.2.3-1build1
pkg|python-webob|1.1.1-1ubuntu0
pkg|python-greenlet|0.3.1-1ubuntu5.1
pkg|python-routes|1.12.3-1ubuntu1
pkg|libldap2-dev|2.4.28-1.1ubuntu4.1
pkg|libsasl2-dev|2.1.25.dfsg1-3ubuntu0.1
pkg|python-bcrypt|0.1-1build2
pkg|gcc|4:4.6.3-1ubuntu5
pkg|python-dateutil|1.5-1
pkg|dnsmasq-base|2.59-4
pkg|dnsmasq-utils|2.59-4
pkg|kpartx|0.4.9-3ubuntu5
pkg|parted|2.3-8ubuntu5
pkg|python-mysqldb|1.2.3-1build1
pkg|python-xattr|0.6.2-1ubuntu1
pkg|python-lxml|2.3.2-1
pkg|kvm|1:84+dfsg-0ubuntu16+1.0+noroms+0ubuntu14.1
pkg|gawk|1:3.1.8+dfsg-0.1ubuntu1
pkg|iptables|1.4.12-1ubuntu4
pkg|ebtables|2.0.9.2-2ubuntu2
pkg|sqlite3|3.7.9-2ubuntu1.1
pkg|sudo|1.8.3p1-1ubuntu3.3
pkg|kvm|1:84+dfsg-0ubuntu16+1.0+noroms+0ubuntu14.1
pkg|vlan|1.9-3ubuntu6
pkg|curl|7.22.0-3ubuntu4
pkg|socat|1.7.1.3-1.2
pkg|python-mox|0.5.3-1ubuntu5
pkg|python-paste|1.7.5.1-4ubuntu2
pkg|python-migrate|0.7.2-1ubuntu1
pkg|python-gflags|1.5.1-1build1
pkg|python-greenlet|0.3.1-1ubuntu5.1
pkg|python-libvirt|0.9.8-2ubuntu17.3
pkg|python-libxml2|2.7.8.dfsg-5.1ubuntu4.1
pkg|python-routes|1.12.3-1ubuntu1
pkg|python-netaddr|0.7.5-4build2
pkg|python-pastedeploy|1.5.0-2build1
pkg|python-eventlet|0.9.16-1ubuntu4.1
pkg|python-cheetah|2.4.4-2ubuntu3
pkg|python-carrot|0.10.7-1ubuntu1
pkg|python-tempita|0.5.1-1build1
pkg|python-sqlalchemy|0.7.4-1ubuntu0.1
pkg|python-suds|0.4.1-2ubuntu1
pkg|python-lockfile|1:0.8-2ubuntu1
pkg|python-m2crypto|0.21.1-2ubuntu2
pkg|python-boto|2.2.2-0ubuntu2
pkg|python-kombu|1.4.3-1
pkg|python-feedparser|5.1-0ubuntu3.1
pkg|python-iso8601|0.1.4-1ubuntu1
pkg|python-qpid|0.12-2
pkg|lvm2|2.02.66-4ubuntu7.1
pkg|open-iscsi|2.0.871-0ubuntu9
pkg|open-iscsi-utils|2.0.871-0ubuntu9
pkg|tgt|1:1.0.17-1ubuntu2
pkg|lvm2|2.02.66-4ubuntu7.1
pkg|python-numpy|1:1.6.1-6ubuntu1
pkg|python-beautifulsoup|3.2.0-2build1
pkg|python-dateutil|1.5-1
pkg|python-paste|1.7.5.1-4ubuntu2
pkg|python-pastedeploy|1.5.0-2build1
pkg|python-anyjson|0.3.1-1build1
pkg|python-routes|1.12.3-1ubuntu1
pkg|python-xattr|0.6.2-1ubuntu1
pkg|python-sqlalchemy|0.7.4-1ubuntu0.1
pkg|python-webob|1.1.1-1ubuntu0
pkg|python-kombu|1.4.3-1
pkg|pylint|0.25.0-1ubuntu1
pkg|pep8|0.6.1-2ubuntu2
pkg|python-eventlet|0.9.16-1ubuntu4.1
pkg|python-nose|1.1.2-3
pkg|python-sphinx|1.1.3+dfsg-2ubuntu2.1
pkg|python-mox|0.5.3-1ubuntu5
pkg|python-kombu|1.4.3-1
pkg|python-coverage|3.4-1ubuntu1
pkg|python-cherrypy3|3.2.2-2
pkg|python-migrate|0.7.2-1ubuntu1
pkg|nodejs|0.6.12~dfsg1-1ubuntu1
pkg|python-netaddr|0.7.5-4build2
pip|Babel|0.9.6
pip|BeautifulSoup|3.2.0
pip|Cheetah|2.4.4
pip|CherryPy|3.2.2
pip|Django|1.4.1
pip|FormEncode|1.2.4
pip|GnuPGInterface|0.3.2
pip|Jinja2|2.6
pip|M2Crypto|0.21.1
pip|Markdown|2.2.0
pip|MarkupSafe|0.15
pip|MySQL-python|1.2.3
pip|PIL|1.1.7
pip|Paste|1.7.5.1
pip|PasteDeploy|1.5.0
pip|PasteScript|1.7.5
pip|PyYAML|3.10
pip|Pygments|1.4
pip|Routes|1.12.3
pip|SQLAlchemy|0.7.4
pip|Sphinx|1.1.3
pip|Tempita|0.5.1
pip|WebOb|1.0.8
pip|amqplib|0.6.1
pip|anyjson|0.2.4
pip|apt-xapian-index|0.44
pip|argparse|1.2.1
pip|bcrypt|0.1
pip|boto|2.1.1
pip|carrot|0.10.7
pip|chardet|2.0.1
pip|https://github.com/openstack/cinder|45e67e8b77ffbc783da111dda96add68f1687341
pip|cliff|1.2.1
pip|cmd2|0.6.3
pip|command-not-found|0.2.44
pip|coverage|3.4
pip|decorator|3.3.2
pip|dingus|0.3.2
pip|distribute|0.6.24dev-r0
pip|django-appconf|0.5
pip|django-compressor|1.2
pip|django-openstack-auth|1.0.2
pip|docutils|0.8.1
pip|euca2ools|2.0.0
pip|eventlet|0.9.17
pip|feedparser|5.1
pip|https://github.com/openstack/glance.git|923b901f8588b4db4a732b01f338ffaa799f3913
pip|greenlet|0.3.1
pip|https://github.com/openstack/horizon.git|b8c2634f658cbfb67703fa9e78e794110a25ef4e
pip|httplib2|0.7.6
pip|iso8601|0.1.4
pip|jsonschema|0.2
pip|keyring|0.9.2
pip|https://github.com/openstack/keystone.git|f261f718eebdfc60cdf9ce1f227c43f462dc163c
pip|kombu|1.0.4
pip|language-selector|0.1
pip|lockfile|0.8
pip|logilab-astng|0.23.0
pip|logilab-common|0.57.1
pip|lxml|2.3.2
pip|mox|0.5.3
pip|netaddr|0.7.5
pip|nose|1.1.2
pip|https://github.com/openstack/nova.git|737ca6e7ecdfa9ffd6744e1e1a54c53c4c897d83
pip|numpy|1.6.1
pip|pam|0.1.4
pip|paramiko|1.7.7.1
pip|passlib|1.6.1
pip|pep8|0.6.1
pip|prettytable|0.6.1
pip|pyOpenSSL|0.12
pip|pycrypto|2.4.1
pip|pylint|0.25.0
pip|pyparsing|1.5.6
pip|pysqlite|2.6.3
pip|python-apt|0.8.3ubuntu4
pip|https://github.com/openstack/python-cinderclient|45bcdcaa2288a09c44306a3d0f0245df4d014897
pip|python-dateutil|1.5
pip|python-debian|0.1.21ubuntu1
pip|python-gflags|1.5.1
pip|https://github.com/openstack/python-glanceclient|e140dbb0c779de74e4ae063971660ac2d234365a
pip|https://github.com/openstack/python-keystoneclient|0a8c96073c3dd76df346c262898303f6fb7435d6
pip|https://github.com/openstack/python-novaclient.git|bfb0f70f40fdea6ee7b75dc2fb44baeadceb646d
pip|python-openid|2.2.5
pip|https://github.com/openstack/python-openstackclient.git|cd140091b5db30e9e3c95929cb5024c3b54ab139
pip|python-neutronclient|2.1
pip|python-swiftclient|1.2.0
pip|pytz|2012d
pip|qpid-python|0.12
pip|scgi|1.13
pip|setuptools-git|0.4.2
pip|simplejson|2.6.1
pip|sqlalchemy-migrate|0.7.2
pip|stompy|0.2.9
pip|suds|0.4
pip|uTidylib|0.2
pip|ufw|0.31-0ubuntu2
pip|unittest2|0.5.1
pip|virtualenv|1.7.1.2
pip|warlock|0.4.0
pip|wsgiref|0.1.2
pip|xattr|0.6.2