For more detail, see the doc migration spec. http://specs.openstack.org/openstack/docs-specs/specs/pike/os-manuals-migration.html Co-Authored-By: Eduardo Gonzalez <dabarren@gmail.com> Change-Id: I3a7c0ed204ee1e9060b5325f20622afe9a5e3040
4.6 KiB
Production architecture guide
This guide will help with configuring Kolla to suit production needs. It is meant to answer some questions regarding basic configuration options that Kolla requires. This document also contains other useful pointers.
Node types and services running on them
A basic Kolla inventory consists of several types of nodes, known in
Ansible as groups
.
- Control - Cloud controller nodes which host control services like APIs and databases. This group should have odd number of nodes for quorum.
- Network - Network nodes host Neutron agents along with haproxy /
keepalived. These nodes will have a floating ip defined in
kolla_internal_vip_address
. - Compute - Compute nodes for compute services. This is where guest VMs live.
- Storage - Storage nodes, for cinder-volume, LVM or ceph-osd.
- Monitoring - Monitor nodes which host monitoring services.
Network configuration
Interface configuration
In Kolla operators should configure following network interfaces:
network_interface
- While it is not used on its own, this provides the required default for other interfaces below.api_interface
- This interface is used for the management network. The management network is the network OpenStack services uses to communicate to each other and the databases. There are known security risks here, so it's recommended to make this network internal, not accessible from outside. Defaults tonetwork_interface
.kolla_external_vip_interface
- This interface is public-facing one. It's used when you want HAProxy public endpoints to be exposed in different network than internal ones. It is mandatory to set this option whenkolla_enable_tls_external
is set to yes. Defaults tonetwork_interface
.storage_interface
- This is the interface that is used by virtual machines to communicate to Ceph. This can be heavily utilized so it's recommended to put this network on 10Gig networking. Defaults tonetwork_interface
.cluster_interface
- This is another interface used by Ceph. It's used for data replication. It can be heavily utilized also and if it becomes a bottleneck it can affect data consistency and performance of whole cluster. Defaults tonetwork_interface
.tunnel_interface
- This interface is used by Neutron for vm-to-vm traffic over tunneled networks (like VxLan). Defaults tonetwork_interface
.neutron_external_interface
- This interface is required by Neutron. Neutron will put br-ex on it. It will be used for flat networking as well as tagged vlan networks. Has to be set separately.dns_interface
- This interface is required by Designate and Bind9. Is used by public facing DNS requests and queries to bind9 and designate mDNS services. Defaults tonetwork_interface
.bifrost_network_interface
- This interface is required by Bifrost. Is used to provision bare metal cloud hosts, require L2 connectivity with the bare metal cloud hosts in order to provide DHCP leases with PXE boot options. Defaults tonetwork_interface
.
Warning
Ansible facts does not recognize interface names containing dashes,
in example br-ex
or bond-0
cannot be used
because ansible will read them as br_ex
and
bond_0
respectively.
Docker configuration
Because Docker is core dependency of Kolla, proper configuration of Docker can change the experience of Kolla significantly. Following section will highlight several Docker configuration details relevant to Kolla operators.
Storage driver
In certain distributions Docker storage driver defaults to devicemapper, which can heavily hit performance of builds and deploys. We suggest to use btrfs or aufs as driver. More details on which storage driver to use in Docker documentation.
Volumes
Kolla puts nearly all of persistent data in Docker volumes. These volumes are created in Docker working directory, which defaults to
/var/lib/docker
We recommend to ensure that this directory has enough space and is placed on fast disk as it will affect performance of builds, deploys as well as database commits and rabbitmq.
This becomes especially relevant when
enable_central_logging
and
openstack_logging_debug
are both set to true, as fully
loaded 130 node cluster produced 30-50GB of logs daily.