Fix the following minor problems to reduce work after stable/liberty branching: 1) RDO: Revert Python MySQL library from PyMySQL to MySQL-python due to lack of support for the former. 2) RDO: Explicitly install 'ebtables' and 'ipset' packages due to dependency problems. 3) General: Change numbered list to bulleted list for lists with only one item. 4) General: Restructure horizon content to match other services. More duplication of content, but sometimes RST conditionals are terrible and distro packages should use the same configuration files. 5) General: Restructure NoSQL content to match SQL content. 6) General: Improve clarity of NTP content. Change-Id: I2620250aa27c7d41b525aa2646ad25e0692140c4 Closes-Bug: #1514760 Closes-Bug: #1514683 Implements: bp installguide-liberty
12 KiB
Networking Option 2: Self-service networks
Install and configure the Networking components on the controller node.
Install the components
ubuntu
# apt-get install neutron-server neutron-plugin-ml2 \
neutron-plugin-linuxbridge-agent neutron-l3-agent neutron-dhcp-agent \
neutron-metadata-agent python-neutronclient
rdo
# yum install openstack-neutron openstack-neutron-ml2 \
openstack-neutron-linuxbridge python-neutronclient ebtables ipset
obs
# zypper install --no-recommends openstack-neutron \
openstack-neutron-server openstack-neutron-linuxbridge-agent \
openstack-neutron-l3-agent openstack-neutron-dhcp-agent \
openstack-neutron-metadata-agent ipset
debian
Install and configure the Networking components
# apt-get install neutron-server neutron-plugin-linuxbridge-agent \ neutron-dhcp-agent neutron-metadata-agent
For networking option 2, also install the
neutron-l3-agent
package.Respond to prompts for database management, Identity service credentials, service endpoint registration, and message queue credentials.
Select the ML2 plug-in:
Note
Selecting the ML2 plug-in also populates the
service_plugins
andallow_overlapping_ips
options in the/etc/neutron/neutron.conf
file with the appropriate values.
ubuntu or rdo or obs
Configure the server component
Edit the
/etc/neutron/neutron.conf
file and complete the following actions:In the
[database]
section, configure database access:ubuntu or obs
[database] ... connection = mysql+pymysql://neutron:NEUTRON_DBPASS@controller/neutron
rdo
[database] ... connection = mysql://neutron:NEUTRON_DBPASS@controller/neutron
Replace
NEUTRON_DBPASS
with the password you chose for the database.In the
[DEFAULT]
section, enable the Modular Layer 2 (ML2) plug-in, router service, and overlapping IP addresses:[DEFAULT] ... core_plugin = ml2 service_plugins = router allow_overlapping_ips = True
In the
[DEFAULT]
and[oslo_messaging_rabbit]
sections, configure RabbitMQ message queue access:[DEFAULT] ... rpc_backend = rabbit [oslo_messaging_rabbit] ... rabbit_host = controller rabbit_userid = openstack rabbit_password = RABBIT_PASS
Replace
RABBIT_PASS
with the password you chose for theopenstack
account in RabbitMQ.In the
[DEFAULT]
and[keystone_authtoken]
sections, configure Identity service access:[DEFAULT] ... auth_strategy = keystone [keystone_authtoken] ... auth_uri = http://controller:5000 auth_url = http://controller:35357 auth_plugin = password project_domain_id = default user_domain_id = default project_name = service username = neutron password = NEUTRON_PASS
Replace
NEUTRON_PASS
with the password you chose for theneutron
user in the Identity service.Note
Comment out or remove any other options in the
[keystone_authtoken]
section.In the
[DEFAULT]
and[nova]
sections, configure Networking to notify Compute of network topology changes:[DEFAULT] ... notify_nova_on_port_status_changes = True notify_nova_on_port_data_changes = True nova_url = http://controller:8774/v2 [nova] ... auth_url = http://controller:35357 auth_plugin = password project_domain_id = default user_domain_id = default region_name = RegionOne project_name = service username = nova password = NOVA_PASS
Replace
NOVA_PASS
with the password you chose for thenova
user in the Identity service.
rdo
In the
[oslo_concurrency]
section, configure the lock path:[oslo_concurrency] ... lock_path = /var/lib/neutron/tmp
(Optional) To assist with troubleshooting, enable verbose logging in the
[DEFAULT]
section:[DEFAULT] ... verbose = True
Configure the Modular Layer 2 (ML2) plug-in
The ML2 plug-in uses the Linux bridge mechanism to build layer-2 (bridging and switching) virtual networking infrastructure for instances.
- Edit the
/etc/neutron/plugins/ml2/ml2_conf.ini
file and complete the following actions:In the
[ml2]
section, enable flat, VLAN, and VXLAN networks:[ml2] ... type_drivers = flat,vlan,vxlan
In the
[ml2]
section, enable VXLAN project (private) networks:[ml2] ... tenant_network_types = vxlan
In the
[ml2]
section, enable the Linux bridge and layer-2 population mechanisms:[ml2] ... mechanism_drivers = linuxbridge,l2population
Warning
After you configure the ML2 plug-in, removing values in the
type_drivers
option can lead to database inconsistency.Note
The Linux bridge agent only supports VXLAN overlay networks.
In the
[ml2]
section, enable the port security extension driver:[ml2] ... extension_drivers = port_security
In the
[ml2_type_flat]
section, configure the public flat provider network:[ml2_type_flat] ... flat_networks = public
In the
[ml2_type_vxlan]
section, configure the VXLAN network identifier range for private networks:[ml2_type_vxlan] ... vni_ranges = 1:1000
In the
[securitygroup]
section, enableipset
to increase efficiency of security group rules:[securitygroup] ... enable_ipset = True
Configure the Linux bridge agent
The Linux bridge agent builds layer-2 (bridging and switching) virtual networking infrastructure for instances including VXLAN tunnels for private networks and handles security groups.
- Edit the
/etc/neutron/plugins/ml2/linuxbridge_agent.ini
file and complete the following actions:In the
[linux_bridge]
section, map the public virtual network to the public physical network interface:[linux_bridge] physical_interface_mappings = public:PUBLIC_INTERFACE_NAME
Replace
PUBLIC_INTERFACE_NAME
with the name of the underlying physical public network interface.In the
[vxlan]
section, enable VXLAN overlay networks, configure the IP address of the physical network interface that handles overlay networks, and enable layer-2 population:[vxlan] enable_vxlan = True local_ip = OVERLAY_INTERFACE_IP_ADDRESS l2_population = True
Replace
OVERLAY_INTERFACE_IP_ADDRESS
with the IP address of the underlying physical network interface that handles overlay networks. The example architecture uses the management interface.In the
[agent]
section, enable ARP spoofing protection:[agent] ... prevent_arp_spoofing = True
In the
[securitygroup]
section, enable security groups and configure the Linux bridgeiptables
firewall driver:[securitygroup] ... enable_security_group = True firewall_driver = neutron.agent.linux.iptables_firewall.IptablesFirewallDriver
Configure the layer-3 agent
The Layer-3 (L3) agent
provides routing and NAT services
for virtual networks.
- Edit the
/etc/neutron/l3_agent.ini
file and complete the following actions:In the
[DEFAULT]
section, configure the Linux bridge interface driver and external network bridge:[DEFAULT] ... interface_driver = neutron.agent.linux.interface.BridgeInterfaceDriver external_network_bridge =
Note
The
external_network_bridge
option intentionally lacks a value to enable multiple external networks on a single agent.(Optional) To assist with troubleshooting, enable verbose logging in the
[DEFAULT]
section:[DEFAULT] ... verbose = True
Configure the DHCP agent
The DHCP agent
provides DHCP services for virtual networks.
Edit the
/etc/neutron/dhcp_agent.ini
file and complete the following actions:In the
[DEFAULT]
section, configure the Linux bridge interface driver, Dnsmasq DHCP driver, and enable isolated metadata so instances on public networks can access metadata over the network:[DEFAULT] ... interface_driver = neutron.agent.linux.interface.BridgeInterfaceDriver dhcp_driver = neutron.agent.linux.dhcp.Dnsmasq enable_isolated_metadata = True
(Optional) To assist with troubleshooting, enable verbose logging in the
[DEFAULT]
section:[DEFAULT] ... verbose = True
Overlay networks such as VXLAN include additional packet headers that increase overhead and decrease space available for the payload or user data. Without knowledge of the virtual network infrastructure, instances attempt to send packets using the default Ethernet
maximum transmission unit (MTU)
of 1500 bytes.Internet protocol (IP)
networks contain thepath MTU discovery (PMTUD)
mechanism to detect end-to-end MTU and adjust packet size accordingly. However, some operating systems and networks block or otherwise lack support for PMTUD causing performance degradation or connectivity failure.Ideally, you can prevent these problems by enabling
jumbo frames <jumbo frame>
on the physical network that contains your tenant virtual networks. Jumbo frames support MTUs up to approximately 9000 bytes which negates the impact of VXLAN overhead on virtual networks. However, many network devices lack support for jumbo frames and OpenStack administrators often lack control over network infrastructure. Given the latter complications, you can also prevent MTU problems by reducing the instance MTU to account for VXLAN overhead. Determining the proper MTU value often takes experimentation, but 1450 bytes works in most environments. You can configure the DHCP server that assigns IP addresses to your instances to also adjust the MTU.Note
Some cloud images ignore the DHCP MTU option in which case you should configure it using metadata, a script, or other suitable method.
In the
[DEFAULT]
section, enable thednsmasq
configuration file:[DEFAULT] ... dnsmasq_config_file = /etc/neutron/dnsmasq-neutron.conf
Create and edit the
/etc/neutron/dnsmasq-neutron.conf
file to enable the DHCP MTU option (26) and configure it to 1450 bytes:dhcp-option-force=26,1450
Return to Networking controller node configuration
<neutron-controller-metadata-agent>
.