ece6a332b7
* Optimize loops * Move install steps to the lib/swift Change-Id: Ie8a74b2627395620ccb0501171fa0150ee7497f2
1743 lines
61 KiB
Bash
Executable File
1743 lines
61 KiB
Bash
Executable File
#!/usr/bin/env bash
|
|
|
|
# ``stack.sh`` is an opinionated OpenStack developer installation. It
|
|
# installs and configures various combinations of **Ceilometer**, **Cinder**,
|
|
# **Glance**, **Heat**, **Horizon**, **Keystone**, **Nova**, **Quantum**
|
|
# and **Swift**
|
|
|
|
# This script allows you to specify configuration options of what git
|
|
# repositories to use, enabled services, network configuration and various
|
|
# passwords. If you are crafty you can run the script on multiple nodes using
|
|
# shared settings for common resources (mysql, rabbitmq) and build a multi-node
|
|
# developer install.
|
|
|
|
# To keep this script simple we assume you are running on a recent **Ubuntu**
|
|
# (11.10 Oneiric or newer) or **Fedora** (F16 or newer) machine. It
|
|
# should work in a VM or physical server. Additionally we put the list of
|
|
# ``apt`` and ``rpm`` dependencies and other configuration files in this repo.
|
|
|
|
# Learn more and get the most recent version at http://devstack.org
|
|
|
|
# Keep track of the devstack directory
|
|
TOP_DIR=$(cd $(dirname "$0") && pwd)
|
|
|
|
# Import common functions
|
|
source $TOP_DIR/functions
|
|
|
|
# Determine what system we are running on. This provides ``os_VENDOR``,
|
|
# ``os_RELEASE``, ``os_UPDATE``, ``os_PACKAGE``, ``os_CODENAME``
|
|
# and ``DISTRO``
|
|
GetDistro
|
|
|
|
|
|
|
|
# Settings
|
|
# ========
|
|
|
|
# ``stack.sh`` is customizable through setting environment variables. If you
|
|
# want to override a setting you can set and export it::
|
|
#
|
|
# export DATABASE_PASSWORD=anothersecret
|
|
# ./stack.sh
|
|
#
|
|
# You can also pass options on a single line ``DATABASE_PASSWORD=simple ./stack.sh``
|
|
#
|
|
# Additionally, you can put any local variables into a ``localrc`` file::
|
|
#
|
|
# DATABASE_PASSWORD=anothersecret
|
|
# DATABASE_USER=hellaroot
|
|
#
|
|
# We try to have sensible defaults, so you should be able to run ``./stack.sh``
|
|
# in most cases. ``localrc`` is not distributed with DevStack and will never
|
|
# be overwritten by a DevStack update.
|
|
#
|
|
# DevStack distributes ``stackrc`` which contains locations for the OpenStack
|
|
# repositories and branches to configure. ``stackrc`` sources ``localrc`` to
|
|
# allow you to safely override those settings.
|
|
|
|
if [[ ! -r $TOP_DIR/stackrc ]]; then
|
|
echo "ERROR: missing $TOP_DIR/stackrc - did you grab more than just stack.sh?"
|
|
exit 1
|
|
fi
|
|
source $TOP_DIR/stackrc
|
|
|
|
|
|
# Proxy Settings
|
|
# --------------
|
|
|
|
# HTTP and HTTPS proxy servers are supported via the usual environment variables [1]
|
|
# ``http_proxy``, ``https_proxy`` and ``no_proxy``. They can be set in
|
|
# ``localrc`` if necessary or on the command line::
|
|
#
|
|
# [1] http://www.w3.org/Daemon/User/Proxies/ProxyClients.html
|
|
#
|
|
# http_proxy=http://proxy.example.com:3128/ no_proxy=repo.example.net ./stack.sh
|
|
|
|
if [[ -n "$http_proxy" ]]; then
|
|
export http_proxy=$http_proxy
|
|
fi
|
|
if [[ -n "$https_proxy" ]]; then
|
|
export https_proxy=$https_proxy
|
|
fi
|
|
if [[ -n "$no_proxy" ]]; then
|
|
export no_proxy=$no_proxy
|
|
fi
|
|
|
|
# Destination path for installation ``DEST``
|
|
DEST=${DEST:-/opt/stack}
|
|
|
|
|
|
# Sanity Check
|
|
# ============
|
|
|
|
# Import database configuration
|
|
source $TOP_DIR/lib/database
|
|
|
|
# Validate database selection
|
|
# Since DATABASE_BACKENDS is now set, this also gets ENABLED_SERVICES
|
|
# properly configured for the database selection.
|
|
use_database $DATABASE_TYPE || echo "Invalid database '$DATABASE_TYPE'"
|
|
|
|
# Remove services which were negated in ENABLED_SERVICES
|
|
# using the "-" prefix (e.g., "-rabbit") instead of
|
|
# calling disable_service().
|
|
disable_negated_services
|
|
|
|
# Warn users who aren't on an explicitly supported distro, but allow them to
|
|
# override check and attempt installation with ``FORCE=yes ./stack``
|
|
if [[ ! ${DISTRO} =~ (oneiric|precise|quantal|raring|f16|f17|f18) ]]; then
|
|
echo "WARNING: this script has not been tested on $DISTRO"
|
|
if [[ "$FORCE" != "yes" ]]; then
|
|
echo "If you wish to run this script anyway run with FORCE=yes"
|
|
exit 1
|
|
fi
|
|
fi
|
|
|
|
if is_service_enabled qpid && ! qpid_is_supported; then
|
|
echo "Qpid support is not available for this version of your distribution."
|
|
exit 1
|
|
fi
|
|
|
|
# ``stack.sh`` keeps function libraries here
|
|
# Make sure ``$TOP_DIR/lib`` directory is present
|
|
if [ ! -d $TOP_DIR/lib ]; then
|
|
echo "ERROR: missing devstack/lib"
|
|
exit 1
|
|
fi
|
|
|
|
# ``stack.sh`` keeps the list of ``apt`` and ``rpm`` dependencies and config
|
|
# templates and other useful files in the ``files`` subdirectory
|
|
FILES=$TOP_DIR/files
|
|
if [ ! -d $FILES ]; then
|
|
echo "ERROR: missing devstack/files"
|
|
exit 1
|
|
fi
|
|
|
|
SCREEN_NAME=${SCREEN_NAME:-stack}
|
|
# Check to see if we are already running DevStack
|
|
if type -p screen >/dev/null && screen -ls | egrep -q "[0-9].$SCREEN_NAME"; then
|
|
echo "You are already running a stack.sh session."
|
|
echo "To rejoin this session type 'screen -x stack'."
|
|
echo "To destroy this session, type './unstack.sh'."
|
|
exit 1
|
|
fi
|
|
|
|
# Make sure we only have one rpc backend enabled.
|
|
rpc_backend_cnt=0
|
|
for svc in qpid zeromq rabbit; do
|
|
is_service_enabled $svc &&
|
|
((rpc_backend_cnt++))
|
|
done
|
|
if [ "$rpc_backend_cnt" -gt 1 ]; then
|
|
echo "ERROR: only one rpc backend may be enabled,"
|
|
echo " set only one of 'rabbit', 'qpid', 'zeromq'"
|
|
echo " via ENABLED_SERVICES."
|
|
elif [ "$rpc_backend_cnt" == 0 ]; then
|
|
echo "ERROR: at least one rpc backend must be enabled,"
|
|
echo " set one of 'rabbit', 'qpid', 'zeromq'"
|
|
echo " via ENABLED_SERVICES."
|
|
fi
|
|
unset rpc_backend_cnt
|
|
|
|
# Set up logging level
|
|
VERBOSE=$(trueorfalse True $VERBOSE)
|
|
|
|
|
|
# root Access
|
|
# -----------
|
|
|
|
# OpenStack is designed to be run as a non-root user; Horizon will fail to run
|
|
# as **root** since Apache will not serve content from **root** user). If
|
|
# ``stack.sh`` is run as **root**, it automatically creates a **stack** user with
|
|
# sudo privileges and runs as that user.
|
|
|
|
if [[ $EUID -eq 0 ]]; then
|
|
ROOTSLEEP=${ROOTSLEEP:-10}
|
|
echo "You are running this script as root."
|
|
echo "In $ROOTSLEEP seconds, we will create a user 'stack' and run as that user"
|
|
sleep $ROOTSLEEP
|
|
|
|
# Give the non-root user the ability to run as **root** via ``sudo``
|
|
is_package_installed sudo || install_package sudo
|
|
if ! getent group stack >/dev/null; then
|
|
echo "Creating a group called stack"
|
|
groupadd stack
|
|
fi
|
|
if ! getent passwd stack >/dev/null; then
|
|
echo "Creating a user called stack"
|
|
useradd -g stack -s /bin/bash -d $DEST -m stack
|
|
fi
|
|
|
|
echo "Giving stack user passwordless sudo privileges"
|
|
# UEC images ``/etc/sudoers`` does not have a ``#includedir``, add one
|
|
grep -q "^#includedir.*/etc/sudoers.d" /etc/sudoers ||
|
|
echo "#includedir /etc/sudoers.d" >> /etc/sudoers
|
|
( umask 226 && echo "stack ALL=(ALL) NOPASSWD:ALL" \
|
|
> /etc/sudoers.d/50_stack_sh )
|
|
|
|
echo "Copying files to stack user"
|
|
STACK_DIR="$DEST/${PWD##*/}"
|
|
cp -r -f -T "$PWD" "$STACK_DIR"
|
|
chown -R stack "$STACK_DIR"
|
|
if [[ "$SHELL_AFTER_RUN" != "no" ]]; then
|
|
exec su -c "set -e; cd $STACK_DIR; bash stack.sh; bash" stack
|
|
else
|
|
exec su -c "set -e; cd $STACK_DIR; bash stack.sh" stack
|
|
fi
|
|
exit 1
|
|
else
|
|
# We're not **root**, make sure ``sudo`` is available
|
|
is_package_installed sudo || die "Sudo is required. Re-run stack.sh as root ONE TIME ONLY to set up sudo."
|
|
|
|
# UEC images ``/etc/sudoers`` does not have a ``#includedir``, add one
|
|
sudo grep -q "^#includedir.*/etc/sudoers.d" /etc/sudoers ||
|
|
echo "#includedir /etc/sudoers.d" | sudo tee -a /etc/sudoers
|
|
|
|
# Set up devstack sudoers
|
|
TEMPFILE=`mktemp`
|
|
echo "`whoami` ALL=(root) NOPASSWD:ALL" >$TEMPFILE
|
|
# Some binaries might be under /sbin or /usr/sbin, so make sure sudo will
|
|
# see them by forcing PATH
|
|
echo "Defaults:`whoami` secure_path=/sbin:/usr/sbin:/usr/bin:/bin:/usr/local/sbin:/usr/local/bin" >> $TEMPFILE
|
|
chmod 0440 $TEMPFILE
|
|
sudo chown root:root $TEMPFILE
|
|
sudo mv $TEMPFILE /etc/sudoers.d/50_stack_sh
|
|
|
|
# Remove old file
|
|
sudo rm -f /etc/sudoers.d/stack_sh_nova
|
|
fi
|
|
|
|
# Create the destination directory and ensure it is writable by the user
|
|
sudo mkdir -p $DEST
|
|
if [ ! -w $DEST ]; then
|
|
sudo chown `whoami` $DEST
|
|
fi
|
|
|
|
# Set ``OFFLINE`` to ``True`` to configure ``stack.sh`` to run cleanly without
|
|
# Internet access. ``stack.sh`` must have been previously run with Internet
|
|
# access to install prerequisites and fetch repositories.
|
|
OFFLINE=`trueorfalse False $OFFLINE`
|
|
|
|
# Set ``ERROR_ON_CLONE`` to ``True`` to configure ``stack.sh`` to exit if
|
|
# the destination git repository does not exist during the ``git_clone``
|
|
# operation.
|
|
ERROR_ON_CLONE=`trueorfalse False $ERROR_ON_CLONE`
|
|
|
|
# Destination path for service data
|
|
DATA_DIR=${DATA_DIR:-${DEST}/data}
|
|
sudo mkdir -p $DATA_DIR
|
|
sudo chown `whoami` $DATA_DIR
|
|
|
|
|
|
# Common Configuration
|
|
# ====================
|
|
|
|
# Set fixed and floating range here so we can make sure not to use addresses
|
|
# from either range when attempting to guess the IP to use for the host.
|
|
# Note that setting FIXED_RANGE may be necessary when running DevStack
|
|
# in an OpenStack cloud that uses either of these address ranges internally.
|
|
FLOATING_RANGE=${FLOATING_RANGE:-172.24.4.224/28}
|
|
FIXED_RANGE=${FIXED_RANGE:-10.0.0.0/24}
|
|
FIXED_NETWORK_SIZE=${FIXED_NETWORK_SIZE:-256}
|
|
NETWORK_GATEWAY=${NETWORK_GATEWAY:-10.0.0.1}
|
|
|
|
# Find the interface used for the default route
|
|
HOST_IP_IFACE=${HOST_IP_IFACE:-$(ip route | sed -n '/^default/{ s/.*dev \(\w\+\)\s\+.*/\1/; p; }')}
|
|
# Search for an IP unless an explicit is set by ``HOST_IP`` environment variable
|
|
if [ -z "$HOST_IP" -o "$HOST_IP" == "dhcp" ]; then
|
|
HOST_IP=""
|
|
HOST_IPS=`LC_ALL=C ip -f inet addr show ${HOST_IP_IFACE} | awk '/inet/ {split($2,parts,"/"); print parts[1]}'`
|
|
for IP in $HOST_IPS; do
|
|
# Attempt to filter out IP addresses that are part of the fixed and
|
|
# floating range. Note that this method only works if the ``netaddr``
|
|
# python library is installed. If it is not installed, an error
|
|
# will be printed and the first IP from the interface will be used.
|
|
# If that is not correct set ``HOST_IP`` in ``localrc`` to the correct
|
|
# address.
|
|
if ! (address_in_net $IP $FIXED_RANGE || address_in_net $IP $FLOATING_RANGE); then
|
|
HOST_IP=$IP
|
|
break;
|
|
fi
|
|
done
|
|
if [ "$HOST_IP" == "" ]; then
|
|
echo "Could not determine host ip address."
|
|
echo "Either localrc specified dhcp on ${HOST_IP_IFACE} or defaulted"
|
|
exit 1
|
|
fi
|
|
fi
|
|
|
|
# Allow the use of an alternate hostname (such as localhost/127.0.0.1) for service endpoints.
|
|
SERVICE_HOST=${SERVICE_HOST:-$HOST_IP}
|
|
|
|
# Configure services to use syslog instead of writing to individual log files
|
|
SYSLOG=`trueorfalse False $SYSLOG`
|
|
SYSLOG_HOST=${SYSLOG_HOST:-$HOST_IP}
|
|
SYSLOG_PORT=${SYSLOG_PORT:-516}
|
|
|
|
# Use color for logging output (only available if syslog is not used)
|
|
LOG_COLOR=`trueorfalse True $LOG_COLOR`
|
|
|
|
# Service startup timeout
|
|
SERVICE_TIMEOUT=${SERVICE_TIMEOUT:-60}
|
|
|
|
|
|
# Configure Projects
|
|
# ==================
|
|
|
|
# Get project function libraries
|
|
source $TOP_DIR/lib/horizon
|
|
source $TOP_DIR/lib/keystone
|
|
source $TOP_DIR/lib/glance
|
|
source $TOP_DIR/lib/nova
|
|
source $TOP_DIR/lib/cinder
|
|
source $TOP_DIR/lib/swift
|
|
source $TOP_DIR/lib/ceilometer
|
|
source $TOP_DIR/lib/heat
|
|
source $TOP_DIR/lib/quantum
|
|
source $TOP_DIR/lib/tempest
|
|
|
|
# Set the destination directories for OpenStack projects
|
|
HORIZON_DIR=$DEST/horizon
|
|
OPENSTACKCLIENT_DIR=$DEST/python-openstackclient
|
|
NOVNC_DIR=$DEST/noVNC
|
|
SWIFT3_DIR=$DEST/swift3
|
|
QUANTUM_DIR=$DEST/quantum
|
|
QUANTUM_CLIENT_DIR=$DEST/python-quantumclient
|
|
|
|
# Default Quantum Plugin
|
|
Q_PLUGIN=${Q_PLUGIN:-openvswitch}
|
|
# Default Quantum Port
|
|
Q_PORT=${Q_PORT:-9696}
|
|
# Default Quantum Host
|
|
Q_HOST=${Q_HOST:-$HOST_IP}
|
|
# Which Quantum API nova should use
|
|
# Default admin username
|
|
Q_ADMIN_USERNAME=${Q_ADMIN_USERNAME:-quantum}
|
|
# Default auth strategy
|
|
Q_AUTH_STRATEGY=${Q_AUTH_STRATEGY:-keystone}
|
|
# Use namespace or not
|
|
Q_USE_NAMESPACE=${Q_USE_NAMESPACE:-True}
|
|
Q_USE_ROOTWRAP=${Q_USE_ROOTWRAP:-True}
|
|
# Meta data IP
|
|
Q_META_DATA_IP=${Q_META_DATA_IP:-$HOST_IP}
|
|
# Use quantum-debug command
|
|
Q_USE_DEBUG_COMMAND=${Q_USE_DEBUG_COMMAND:-False}
|
|
|
|
RYU_DIR=$DEST/ryu
|
|
# Ryu API Host
|
|
RYU_API_HOST=${RYU_API_HOST:-127.0.0.1}
|
|
# Ryu API Port
|
|
RYU_API_PORT=${RYU_API_PORT:-8080}
|
|
# Ryu OFP Host
|
|
RYU_OFP_HOST=${RYU_OFP_HOST:-127.0.0.1}
|
|
# Ryu OFP Port
|
|
RYU_OFP_PORT=${RYU_OFP_PORT:-6633}
|
|
# Ryu Applications
|
|
RYU_APPS=${RYU_APPS:-ryu.app.simple_isolation,ryu.app.rest}
|
|
|
|
# Should cinder perform secure deletion of volumes?
|
|
# Defaults to true, can be set to False to avoid this bug when testing:
|
|
# https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1023755
|
|
CINDER_SECURE_DELETE=`trueorfalse True $CINDER_SECURE_DELETE`
|
|
|
|
# Name of the LVM volume group to use/create for iscsi volumes
|
|
VOLUME_GROUP=${VOLUME_GROUP:-stack-volumes}
|
|
VOLUME_NAME_PREFIX=${VOLUME_NAME_PREFIX:-volume-}
|
|
INSTANCE_NAME_PREFIX=${INSTANCE_NAME_PREFIX:-instance-}
|
|
|
|
# Generic helper to configure passwords
|
|
function read_password {
|
|
XTRACE=$(set +o | grep xtrace)
|
|
set +o xtrace
|
|
var=$1; msg=$2
|
|
pw=${!var}
|
|
|
|
localrc=$TOP_DIR/localrc
|
|
|
|
# If the password is not defined yet, proceed to prompt user for a password.
|
|
if [ ! $pw ]; then
|
|
# If there is no localrc file, create one
|
|
if [ ! -e $localrc ]; then
|
|
touch $localrc
|
|
fi
|
|
|
|
# Presumably if we got this far it can only be that our localrc is missing
|
|
# the required password. Prompt user for a password and write to localrc.
|
|
echo ''
|
|
echo '################################################################################'
|
|
echo $msg
|
|
echo '################################################################################'
|
|
echo "This value will be written to your localrc file so you don't have to enter it "
|
|
echo "again. Use only alphanumeric characters."
|
|
echo "If you leave this blank, a random default value will be used."
|
|
pw=" "
|
|
while true; do
|
|
echo "Enter a password now:"
|
|
read -e $var
|
|
pw=${!var}
|
|
[[ "$pw" = "`echo $pw | tr -cd [:alnum:]`" ]] && break
|
|
echo "Invalid chars in password. Try again:"
|
|
done
|
|
if [ ! $pw ]; then
|
|
pw=`openssl rand -hex 10`
|
|
fi
|
|
eval "$var=$pw"
|
|
echo "$var=$pw" >> $localrc
|
|
fi
|
|
$XTRACE
|
|
}
|
|
|
|
|
|
# Nova Network Configuration
|
|
# --------------------------
|
|
|
|
# FIXME: more documentation about why these are important options. Also
|
|
# we should make sure we use the same variable names as the option names.
|
|
|
|
if [ "$VIRT_DRIVER" = 'xenserver' ]; then
|
|
PUBLIC_INTERFACE_DEFAULT=eth3
|
|
# Allow ``build_domU.sh`` to specify the flat network bridge via kernel args
|
|
FLAT_NETWORK_BRIDGE_DEFAULT=$(grep -o 'flat_network_bridge=[[:alnum:]]*' /proc/cmdline | cut -d= -f 2 | sort -u)
|
|
GUEST_INTERFACE_DEFAULT=eth1
|
|
else
|
|
PUBLIC_INTERFACE_DEFAULT=br100
|
|
FLAT_NETWORK_BRIDGE_DEFAULT=br100
|
|
GUEST_INTERFACE_DEFAULT=eth0
|
|
fi
|
|
|
|
PUBLIC_INTERFACE=${PUBLIC_INTERFACE:-$PUBLIC_INTERFACE_DEFAULT}
|
|
NET_MAN=${NET_MAN:-FlatDHCPManager}
|
|
EC2_DMZ_HOST=${EC2_DMZ_HOST:-$SERVICE_HOST}
|
|
FLAT_NETWORK_BRIDGE=${FLAT_NETWORK_BRIDGE:-$FLAT_NETWORK_BRIDGE_DEFAULT}
|
|
VLAN_INTERFACE=${VLAN_INTERFACE:-$GUEST_INTERFACE_DEFAULT}
|
|
|
|
# Test floating pool and range are used for testing. They are defined
|
|
# here until the admin APIs can replace nova-manage
|
|
TEST_FLOATING_POOL=${TEST_FLOATING_POOL:-test}
|
|
TEST_FLOATING_RANGE=${TEST_FLOATING_RANGE:-192.168.253.0/29}
|
|
|
|
# ``MULTI_HOST`` is a mode where each compute node runs its own network node. This
|
|
# allows network operations and routing for a VM to occur on the server that is
|
|
# running the VM - removing a SPOF and bandwidth bottleneck.
|
|
MULTI_HOST=`trueorfalse False $MULTI_HOST`
|
|
|
|
# If you are using the FlatDHCP network mode on multiple hosts, set the
|
|
# ``FLAT_INTERFACE`` variable but make sure that the interface doesn't already
|
|
# have an IP or you risk breaking things.
|
|
#
|
|
# **DHCP Warning**: If your flat interface device uses DHCP, there will be a
|
|
# hiccup while the network is moved from the flat interface to the flat network
|
|
# bridge. This will happen when you launch your first instance. Upon launch
|
|
# you will lose all connectivity to the node, and the VM launch will probably
|
|
# fail.
|
|
#
|
|
# If you are running on a single node and don't need to access the VMs from
|
|
# devices other than that node, you can set ``FLAT_INTERFACE=``
|
|
# This will stop nova from bridging any interfaces into ``FLAT_NETWORK_BRIDGE``.
|
|
FLAT_INTERFACE=${FLAT_INTERFACE-$GUEST_INTERFACE_DEFAULT}
|
|
|
|
## FIXME(ja): should/can we check that FLAT_INTERFACE is sane?
|
|
|
|
|
|
# Quantum Networking
|
|
# ------------------
|
|
|
|
# Make sure that quantum is enabled in ENABLED_SERVICES. If you want
|
|
# to run Quantum on this host, make sure that q-svc is also in
|
|
# ENABLED_SERVICES.
|
|
#
|
|
# If you're planning to use the Quantum openvswitch plugin, set
|
|
# Q_PLUGIN to "openvswitch" and make sure the q-agt service is enabled
|
|
# in ENABLED_SERVICES. If you're planning to use the Quantum
|
|
# linuxbridge plugin, set Q_PLUGIN to "linuxbridge" and make sure the
|
|
# q-agt service is enabled in ENABLED_SERVICES.
|
|
#
|
|
# See "Quantum Network Configuration" below for additional variables
|
|
# that must be set in localrc for connectivity across hosts with
|
|
# Quantum.
|
|
#
|
|
# With Quantum networking the NET_MAN variable is ignored.
|
|
|
|
|
|
# Database Configuration
|
|
# ----------------------
|
|
|
|
# To select between database backends, add a line to localrc like:
|
|
#
|
|
# use_database postgresql
|
|
#
|
|
# The available database backends are defined in the ``DATABASE_BACKENDS``
|
|
# variable defined in stackrc. By default, MySQL is enabled as the database
|
|
# backend.
|
|
|
|
initialize_database_backends && echo "Using $DATABASE_TYPE database backend" || echo "No database enabled"
|
|
|
|
|
|
# RabbitMQ or Qpid
|
|
# --------------------------
|
|
|
|
# Rabbit connection info
|
|
if is_service_enabled rabbit; then
|
|
RABBIT_HOST=${RABBIT_HOST:-localhost}
|
|
read_password RABBIT_PASSWORD "ENTER A PASSWORD TO USE FOR RABBIT."
|
|
fi
|
|
|
|
if is_service_enabled swift; then
|
|
# If we are using swift3, we can default the s3 port to swift instead
|
|
# of nova-objectstore
|
|
if is_service_enabled swift3;then
|
|
S3_SERVICE_PORT=${S3_SERVICE_PORT:-8080}
|
|
fi
|
|
# We only ask for Swift Hash if we have enabled swift service.
|
|
# ``SWIFT_HASH`` is a random unique string for a swift cluster that
|
|
# can never change.
|
|
read_password SWIFT_HASH "ENTER A RANDOM SWIFT HASH."
|
|
fi
|
|
|
|
# Set default port for nova-objectstore
|
|
S3_SERVICE_PORT=${S3_SERVICE_PORT:-3333}
|
|
|
|
|
|
# Keystone
|
|
# --------
|
|
|
|
# The ``SERVICE_TOKEN`` is used to bootstrap the Keystone database. It is
|
|
# just a string and is not a 'real' Keystone token.
|
|
read_password SERVICE_TOKEN "ENTER A SERVICE_TOKEN TO USE FOR THE SERVICE ADMIN TOKEN."
|
|
# Services authenticate to Identity with servicename/``SERVICE_PASSWORD``
|
|
read_password SERVICE_PASSWORD "ENTER A SERVICE_PASSWORD TO USE FOR THE SERVICE AUTHENTICATION."
|
|
# Horizon currently truncates usernames and passwords at 20 characters
|
|
read_password ADMIN_PASSWORD "ENTER A PASSWORD TO USE FOR HORIZON AND KEYSTONE (20 CHARS OR LESS)."
|
|
|
|
# Set the tenant for service accounts in Keystone
|
|
SERVICE_TENANT_NAME=${SERVICE_TENANT_NAME:-service}
|
|
|
|
|
|
# Log files
|
|
# ---------
|
|
|
|
# Draw a spinner so the user knows something is happening
|
|
function spinner() {
|
|
local delay=0.75
|
|
local spinstr='/-\|'
|
|
printf "..." >&3
|
|
while [ true ]; do
|
|
local temp=${spinstr#?}
|
|
printf "[%c]" "$spinstr" >&3
|
|
local spinstr=$temp${spinstr%"$temp"}
|
|
sleep $delay
|
|
printf "\b\b\b" >&3
|
|
done
|
|
}
|
|
|
|
# Echo text to the log file, summary log file and stdout
|
|
# echo_summary "something to say"
|
|
function echo_summary() {
|
|
if [[ -t 3 && "$VERBOSE" != "True" ]]; then
|
|
kill >/dev/null 2>&1 $LAST_SPINNER_PID
|
|
if [ ! -z "$LAST_SPINNER_PID" ]; then
|
|
printf "\b\b\bdone\n" >&3
|
|
fi
|
|
echo -n $@ >&6
|
|
spinner &
|
|
LAST_SPINNER_PID=$!
|
|
else
|
|
echo $@ >&6
|
|
fi
|
|
}
|
|
|
|
# Echo text only to stdout, no log files
|
|
# echo_nolog "something not for the logs"
|
|
function echo_nolog() {
|
|
echo $@ >&3
|
|
}
|
|
|
|
# Set up logging for ``stack.sh``
|
|
# Set ``LOGFILE`` to turn on logging
|
|
# Append '.xxxxxxxx' to the given name to maintain history
|
|
# where 'xxxxxxxx' is a representation of the date the file was created
|
|
if [[ -n "$LOGFILE" || -n "$SCREEN_LOGDIR" ]]; then
|
|
LOGDAYS=${LOGDAYS:-7}
|
|
TIMESTAMP_FORMAT=${TIMESTAMP_FORMAT:-"%F-%H%M%S"}
|
|
CURRENT_LOG_TIME=$(date "+$TIMESTAMP_FORMAT")
|
|
fi
|
|
|
|
if [[ -n "$LOGFILE" ]]; then
|
|
# First clean up old log files. Use the user-specified ``LOGFILE``
|
|
# as the template to search for, appending '.*' to match the date
|
|
# we added on earlier runs.
|
|
LOGDIR=$(dirname "$LOGFILE")
|
|
LOGNAME=$(basename "$LOGFILE")
|
|
mkdir -p $LOGDIR
|
|
find $LOGDIR -maxdepth 1 -name $LOGNAME.\* -mtime +$LOGDAYS -exec rm {} \;
|
|
LOGFILE=$LOGFILE.${CURRENT_LOG_TIME}
|
|
SUMFILE=$LOGFILE.${CURRENT_LOG_TIME}.summary
|
|
|
|
# Redirect output according to config
|
|
|
|
# Copy stdout to fd 3
|
|
exec 3>&1
|
|
if [[ "$VERBOSE" == "True" ]]; then
|
|
# Redirect stdout/stderr to tee to write the log file
|
|
exec 1> >( tee "${LOGFILE}" ) 2>&1
|
|
# Set up a second fd for output
|
|
exec 6> >( tee "${SUMFILE}" )
|
|
else
|
|
# Set fd 1 and 2 to primary logfile
|
|
exec 1> "${LOGFILE}" 2>&1
|
|
# Set fd 6 to summary logfile and stdout
|
|
exec 6> >( tee "${SUMFILE}" /dev/fd/3 )
|
|
fi
|
|
|
|
echo_summary "stack.sh log $LOGFILE"
|
|
# Specified logfile name always links to the most recent log
|
|
ln -sf $LOGFILE $LOGDIR/$LOGNAME
|
|
ln -sf $SUMFILE $LOGDIR/$LOGNAME.summary
|
|
else
|
|
# Set up output redirection without log files
|
|
# Copy stdout to fd 3
|
|
exec 3>&1
|
|
if [[ "$VERBOSE" != "True" ]]; then
|
|
# Throw away stdout and stderr
|
|
exec 1>/dev/null 2>&1
|
|
fi
|
|
# Always send summary fd to original stdout
|
|
exec 6>&3
|
|
fi
|
|
|
|
# Set up logging of screen windows
|
|
# Set ``SCREEN_LOGDIR`` to turn on logging of screen windows to the
|
|
# directory specified in ``SCREEN_LOGDIR``, we will log to the the file
|
|
# ``screen-$SERVICE_NAME-$TIMESTAMP.log`` in that dir and have a link
|
|
# ``screen-$SERVICE_NAME.log`` to the latest log file.
|
|
# Logs are kept for as long specified in ``LOGDAYS``.
|
|
if [[ -n "$SCREEN_LOGDIR" ]]; then
|
|
|
|
# We make sure the directory is created.
|
|
if [[ -d "$SCREEN_LOGDIR" ]]; then
|
|
# We cleanup the old logs
|
|
find $SCREEN_LOGDIR -maxdepth 1 -name screen-\*.log -mtime +$LOGDAYS -exec rm {} \;
|
|
else
|
|
mkdir -p $SCREEN_LOGDIR
|
|
fi
|
|
fi
|
|
|
|
|
|
# Set Up Script Execution
|
|
# -----------------------
|
|
|
|
# Kill background processes on exit
|
|
trap clean EXIT
|
|
clean() {
|
|
local r=$?
|
|
kill >/dev/null 2>&1 $(jobs -p)
|
|
exit $r
|
|
}
|
|
|
|
|
|
# Exit on any errors so that errors don't compound
|
|
trap failed ERR
|
|
failed() {
|
|
local r=$?
|
|
kill >/dev/null 2>&1 $(jobs -p)
|
|
set +o xtrace
|
|
[ -n "$LOGFILE" ] && echo "${0##*/} failed: full log in $LOGFILE"
|
|
exit $r
|
|
}
|
|
|
|
# Print the commands being run so that we can see the command that triggers
|
|
# an error. It is also useful for following along as the install occurs.
|
|
set -o xtrace
|
|
|
|
|
|
# Install Packages
|
|
# ================
|
|
|
|
# OpenStack uses a fair number of other projects.
|
|
|
|
# Install package requirements
|
|
echo_summary "Installing package prerequisites"
|
|
if [[ "$os_PACKAGE" = "deb" ]]; then
|
|
install_package $(get_packages $FILES/apts)
|
|
else
|
|
install_package $(get_packages $FILES/rpms)
|
|
fi
|
|
|
|
if [[ $SYSLOG != "False" ]]; then
|
|
install_package rsyslog-relp
|
|
fi
|
|
|
|
if is_service_enabled rabbit; then
|
|
# Install rabbitmq-server
|
|
# the temp file is necessary due to LP: #878600
|
|
tfile=$(mktemp)
|
|
install_package rabbitmq-server > "$tfile" 2>&1
|
|
cat "$tfile"
|
|
rm -f "$tfile"
|
|
elif is_service_enabled qpid; then
|
|
if [[ "$os_PACKAGE" = "rpm" ]]; then
|
|
install_package qpid-cpp-server-daemon
|
|
else
|
|
install_package qpidd
|
|
fi
|
|
elif is_service_enabled zeromq; then
|
|
if [[ "$os_PACKAGE" = "rpm" ]]; then
|
|
install_package zeromq python-zmq
|
|
else
|
|
install_package libzmq1 python-zmq
|
|
fi
|
|
fi
|
|
|
|
if is_service_enabled $DATABASE_BACKENDS; then
|
|
install_database
|
|
fi
|
|
|
|
if is_service_enabled q-agt; then
|
|
if is_quantum_ovs_base_plugin "$Q_PLUGIN"; then
|
|
# Install deps
|
|
# FIXME add to ``files/apts/quantum``, but don't install if not needed!
|
|
if [[ "$os_PACKAGE" = "deb" ]]; then
|
|
kernel_version=`cat /proc/version | cut -d " " -f3`
|
|
install_package make fakeroot dkms openvswitch-switch openvswitch-datapath-dkms linux-headers-$kernel_version
|
|
else
|
|
### FIXME(dtroyer): Find RPMs for OpenVSwitch
|
|
echo "OpenVSwitch packages need to be located"
|
|
# Fedora does not started OVS by default
|
|
restart_service openvswitch
|
|
fi
|
|
elif [[ "$Q_PLUGIN" = "linuxbridge" ]]; then
|
|
install_package bridge-utils
|
|
fi
|
|
fi
|
|
|
|
TRACK_DEPENDS=${TRACK_DEPENDS:-False}
|
|
|
|
# Install python packages into a virtualenv so that we can track them
|
|
if [[ $TRACK_DEPENDS = True ]] ; then
|
|
echo_summary "Installing Python packages into a virtualenv $DEST/.venv"
|
|
install_package python-virtualenv
|
|
|
|
rm -rf $DEST/.venv
|
|
virtualenv --system-site-packages $DEST/.venv
|
|
source $DEST/.venv/bin/activate
|
|
$DEST/.venv/bin/pip freeze > $DEST/requires-pre-pip
|
|
fi
|
|
|
|
# Install python requirements
|
|
echo_summary "Installing Python prerequisites"
|
|
pip_install $(get_packages $FILES/pips | sort -u)
|
|
|
|
|
|
# Check Out Source
|
|
# ----------------
|
|
|
|
echo_summary "Installing OpenStack project source"
|
|
|
|
# Grab clients first
|
|
install_keystoneclient
|
|
install_glanceclient
|
|
install_novaclient
|
|
# Check out the client libs that are used most
|
|
git_clone $OPENSTACKCLIENT_REPO $OPENSTACKCLIENT_DIR $OPENSTACKCLIENT_BRANCH
|
|
|
|
# glance, swift middleware and nova api needs keystone middleware
|
|
if is_service_enabled key g-api n-api swift; then
|
|
# unified auth system (manages accounts/tokens)
|
|
install_keystone
|
|
fi
|
|
|
|
if is_service_enabled swift; then
|
|
install_swiftclient
|
|
install_swift
|
|
if is_service_enabled swift3; then
|
|
# swift3 middleware to provide S3 emulation to Swift
|
|
git_clone $SWIFT3_REPO $SWIFT3_DIR $SWIFT3_BRANCH
|
|
fi
|
|
fi
|
|
|
|
if is_service_enabled g-api n-api; then
|
|
# image catalog service
|
|
install_glance
|
|
fi
|
|
if is_service_enabled nova; then
|
|
# compute service
|
|
install_nova
|
|
fi
|
|
if is_service_enabled n-novnc; then
|
|
# a websockets/html5 or flash powered VNC console for vm instances
|
|
git_clone $NOVNC_REPO $NOVNC_DIR $NOVNC_BRANCH
|
|
fi
|
|
if is_service_enabled horizon; then
|
|
# dashboard
|
|
install_horizon
|
|
fi
|
|
if is_service_enabled quantum; then
|
|
git_clone $QUANTUM_CLIENT_REPO $QUANTUM_CLIENT_DIR $QUANTUM_CLIENT_BRANCH
|
|
fi
|
|
if is_service_enabled quantum; then
|
|
# quantum
|
|
git_clone $QUANTUM_REPO $QUANTUM_DIR $QUANTUM_BRANCH
|
|
fi
|
|
if is_service_enabled heat; then
|
|
install_heat
|
|
install_heatclient
|
|
fi
|
|
if is_service_enabled cinder; then
|
|
install_cinder
|
|
fi
|
|
if is_service_enabled ceilometer; then
|
|
install_ceilometer
|
|
fi
|
|
if is_service_enabled tempest; then
|
|
install_tempest
|
|
fi
|
|
if is_service_enabled ryu || (is_service_enabled quantum && [[ "$Q_PLUGIN" = "ryu" ]]); then
|
|
git_clone $RYU_REPO $RYU_DIR $RYU_BRANCH
|
|
fi
|
|
|
|
|
|
# Initialization
|
|
# ==============
|
|
|
|
echo_summary "Configuring OpenStack projects"
|
|
|
|
# Set up our checkouts so they are installed into python path
|
|
# allowing ``import nova`` or ``import glance.client``
|
|
configure_keystoneclient
|
|
configure_novaclient
|
|
setup_develop $OPENSTACKCLIENT_DIR
|
|
if is_service_enabled key g-api n-api swift; then
|
|
configure_keystone
|
|
fi
|
|
if is_service_enabled swift; then
|
|
configure_swift
|
|
configure_swiftclient
|
|
if is_service_enabled swift3; then
|
|
setup_develop $SWIFT3_DIR
|
|
fi
|
|
fi
|
|
if is_service_enabled g-api n-api; then
|
|
configure_glance
|
|
fi
|
|
|
|
# Do this _after_ glance is installed to override the old binary
|
|
# TODO(dtroyer): figure out when this is no longer necessary
|
|
configure_glanceclient
|
|
|
|
if is_service_enabled nova; then
|
|
configure_nova
|
|
fi
|
|
if is_service_enabled horizon; then
|
|
configure_horizon
|
|
fi
|
|
if is_service_enabled quantum; then
|
|
setup_develop $QUANTUM_CLIENT_DIR
|
|
setup_develop $QUANTUM_DIR
|
|
fi
|
|
if is_service_enabled heat; then
|
|
configure_heat
|
|
configure_heatclient
|
|
fi
|
|
if is_service_enabled cinder; then
|
|
configure_cinder
|
|
fi
|
|
if is_service_enabled ryu || (is_service_enabled quantum && [[ "$Q_PLUGIN" = "ryu" ]]); then
|
|
setup_develop $RYU_DIR
|
|
fi
|
|
|
|
if [[ $TRACK_DEPENDS = True ]] ; then
|
|
$DEST/.venv/bin/pip freeze > $DEST/requires-post-pip
|
|
if ! diff -Nru $DEST/requires-pre-pip $DEST/requires-post-pip > $DEST/requires.diff ; then
|
|
cat $DEST/requires.diff
|
|
fi
|
|
echo "Ran stack.sh in depend tracking mode, bailing out now"
|
|
exit 0
|
|
fi
|
|
|
|
|
|
# Syslog
|
|
# ------
|
|
|
|
if [[ $SYSLOG != "False" ]]; then
|
|
if [[ "$SYSLOG_HOST" = "$HOST_IP" ]]; then
|
|
# Configure the master host to receive
|
|
cat <<EOF >/tmp/90-stack-m.conf
|
|
\$ModLoad imrelp
|
|
\$InputRELPServerRun $SYSLOG_PORT
|
|
EOF
|
|
sudo mv /tmp/90-stack-m.conf /etc/rsyslog.d
|
|
else
|
|
# Set rsyslog to send to remote host
|
|
cat <<EOF >/tmp/90-stack-s.conf
|
|
*.* :omrelp:$SYSLOG_HOST:$SYSLOG_PORT
|
|
EOF
|
|
sudo mv /tmp/90-stack-s.conf /etc/rsyslog.d
|
|
fi
|
|
echo_summary "Starting rsyslog"
|
|
restart_service rsyslog
|
|
fi
|
|
|
|
|
|
# Finalize queue installation
|
|
# ----------------------------
|
|
|
|
if is_service_enabled rabbit; then
|
|
# Start rabbitmq-server
|
|
echo_summary "Starting RabbitMQ"
|
|
if [[ "$os_PACKAGE" = "rpm" ]]; then
|
|
# RPM doesn't start the service
|
|
restart_service rabbitmq-server
|
|
fi
|
|
# change the rabbit password since the default is "guest"
|
|
sudo rabbitmqctl change_password guest $RABBIT_PASSWORD
|
|
elif is_service_enabled qpid; then
|
|
echo_summary "Starting qpid"
|
|
restart_service qpidd
|
|
fi
|
|
|
|
|
|
# Configure database
|
|
# ------------------
|
|
|
|
if is_service_enabled $DATABASE_BACKENDS; then
|
|
configure_database
|
|
fi
|
|
|
|
|
|
# Configure screen
|
|
# ----------------
|
|
|
|
if [ -z "$SCREEN_HARDSTATUS" ]; then
|
|
SCREEN_HARDSTATUS='%{= .} %-Lw%{= .}%> %n%f %t*%{= .}%+Lw%< %-=%{g}(%{d}%H/%l%{g})'
|
|
fi
|
|
|
|
# Clear screen rc file
|
|
SCREENRC=$TOP_DIR/$SCREEN_NAME-screenrc
|
|
if [[ -e $SCREENRC ]]; then
|
|
echo -n > $SCREENRC
|
|
fi
|
|
|
|
# Create a new named screen to run processes in
|
|
screen -d -m -S $SCREEN_NAME -t shell -s /bin/bash
|
|
sleep 1
|
|
|
|
# Set a reasonable status bar
|
|
screen -r $SCREEN_NAME -X hardstatus alwayslastline "$SCREEN_HARDSTATUS"
|
|
|
|
|
|
# Keystone
|
|
# --------
|
|
|
|
if is_service_enabled key; then
|
|
echo_summary "Starting Keystone"
|
|
configure_keystone
|
|
init_keystone
|
|
start_keystone
|
|
echo "Waiting for keystone to start..."
|
|
if ! timeout $SERVICE_TIMEOUT sh -c "while ! http_proxy= curl -s $KEYSTONE_AUTH_PROTOCOL://$SERVICE_HOST:$KEYSTONE_API_PORT/v2.0/ >/dev/null; do sleep 1; done"; then
|
|
echo "keystone did not start"
|
|
exit 1
|
|
fi
|
|
|
|
# ``keystone_data.sh`` creates services, admin and demo users, and roles.
|
|
SERVICE_ENDPOINT=$KEYSTONE_AUTH_PROTOCOL://$KEYSTONE_AUTH_HOST:$KEYSTONE_AUTH_PORT/v2.0
|
|
|
|
ADMIN_PASSWORD=$ADMIN_PASSWORD SERVICE_TENANT_NAME=$SERVICE_TENANT_NAME SERVICE_PASSWORD=$SERVICE_PASSWORD \
|
|
SERVICE_TOKEN=$SERVICE_TOKEN SERVICE_ENDPOINT=$SERVICE_ENDPOINT SERVICE_HOST=$SERVICE_HOST \
|
|
S3_SERVICE_PORT=$S3_SERVICE_PORT KEYSTONE_CATALOG_BACKEND=$KEYSTONE_CATALOG_BACKEND \
|
|
DEVSTACK_DIR=$TOP_DIR ENABLED_SERVICES=$ENABLED_SERVICES HEAT_API_CFN_PORT=$HEAT_API_CFN_PORT \
|
|
HEAT_API_PORT=$HEAT_API_PORT \
|
|
bash -x $FILES/keystone_data.sh
|
|
|
|
# Set up auth creds now that keystone is bootstrapped
|
|
export OS_AUTH_URL=$SERVICE_ENDPOINT
|
|
export OS_TENANT_NAME=admin
|
|
export OS_USERNAME=admin
|
|
export OS_PASSWORD=$ADMIN_PASSWORD
|
|
fi
|
|
|
|
|
|
# Horizon
|
|
# -------
|
|
|
|
# Set up the django horizon application to serve via apache/wsgi
|
|
|
|
if is_service_enabled horizon; then
|
|
echo_summary "Configuring and starting Horizon"
|
|
init_horizon
|
|
start_horizon
|
|
fi
|
|
|
|
|
|
# Glance
|
|
# ------
|
|
|
|
if is_service_enabled g-reg; then
|
|
echo_summary "Configuring Glance"
|
|
|
|
init_glance
|
|
|
|
# Store the images in swift if enabled.
|
|
if is_service_enabled swift; then
|
|
iniset $GLANCE_API_CONF DEFAULT default_store swift
|
|
iniset $GLANCE_API_CONF DEFAULT swift_store_auth_address $KEYSTONE_SERVICE_PROTOCOL://$KEYSTONE_SERVICE_HOST:$KEYSTONE_SERVICE_PORT/v2.0/
|
|
iniset $GLANCE_API_CONF DEFAULT swift_store_user $SERVICE_TENANT_NAME:glance
|
|
iniset $GLANCE_API_CONF DEFAULT swift_store_key $SERVICE_PASSWORD
|
|
iniset $GLANCE_API_CONF DEFAULT swift_store_create_container_on_put True
|
|
fi
|
|
fi
|
|
|
|
|
|
# Ryu
|
|
# ---
|
|
|
|
# Ryu is not a part of OpenStack project. Please ignore following block if
|
|
# you are not interested in Ryu.
|
|
# launch ryu manager
|
|
if is_service_enabled ryu; then
|
|
RYU_CONF_DIR=/etc/ryu
|
|
if [[ ! -d $RYU_CONF_DIR ]]; then
|
|
sudo mkdir -p $RYU_CONF_DIR
|
|
fi
|
|
sudo chown `whoami` $RYU_CONF_DIR
|
|
RYU_CONF=$RYU_CONF_DIR/ryu.conf
|
|
sudo rm -rf $RYU_CONF
|
|
|
|
cat <<EOF > $RYU_CONF
|
|
--app_lists=$RYU_APPS
|
|
--wsapi_host=$RYU_API_HOST
|
|
--wsapi_port=$RYU_API_PORT
|
|
--ofp_listen_host=$RYU_OFP_HOST
|
|
--ofp_tcp_listen_port=$RYU_OFP_PORT
|
|
EOF
|
|
screen_it ryu "cd $RYU_DIR && $RYU_DIR/bin/ryu-manager --flagfile $RYU_CONF"
|
|
fi
|
|
|
|
|
|
# Quantum
|
|
# -------
|
|
|
|
# Quantum Network Configuration
|
|
if is_service_enabled quantum; then
|
|
echo_summary "Configuring Quantum"
|
|
|
|
# The following variables control the Quantum openvswitch and
|
|
# linuxbridge plugins' allocation of tenant networks and
|
|
# availability of provider networks. If these are not configured
|
|
# in localrc, tenant networks will be local to the host (with no
|
|
# remote connectivity), and no physical resources will be
|
|
# available for the allocation of provider networks.
|
|
|
|
# To use GRE tunnels for tenant networks, set to True in
|
|
# localrc. GRE tunnels are only supported by the openvswitch
|
|
# plugin, and currently only on Ubuntu.
|
|
ENABLE_TENANT_TUNNELS=${ENABLE_TENANT_TUNNELS:-False}
|
|
|
|
# If using GRE tunnels for tenant networks, specify the range of
|
|
# tunnel IDs from which tenant networks are allocated. Can be
|
|
# overriden in localrc in necesssary.
|
|
TENANT_TUNNEL_RANGES=${TENANT_TUNNEL_RANGE:-1:1000}
|
|
|
|
# To use VLANs for tenant networks, set to True in localrc. VLANs
|
|
# are supported by the openvswitch and linuxbridge plugins, each
|
|
# requiring additional configuration described below.
|
|
ENABLE_TENANT_VLANS=${ENABLE_TENANT_VLANS:-False}
|
|
|
|
# If using VLANs for tenant networks, set in localrc to specify
|
|
# the range of VLAN VIDs from which tenant networks are
|
|
# allocated. An external network switch must be configured to
|
|
# trunk these VLANs between hosts for multi-host connectivity.
|
|
#
|
|
# Example: ``TENANT_VLAN_RANGE=1000:1999``
|
|
TENANT_VLAN_RANGE=${TENANT_VLAN_RANGE:-}
|
|
|
|
# If using VLANs for tenant networks, or if using flat or VLAN
|
|
# provider networks, set in localrc to the name of the physical
|
|
# network, and also configure OVS_PHYSICAL_BRIDGE for the
|
|
# openvswitch agent or LB_PHYSICAL_INTERFACE for the linuxbridge
|
|
# agent, as described below.
|
|
#
|
|
# Example: ``PHYSICAL_NETWORK=default``
|
|
PHYSICAL_NETWORK=${PHYSICAL_NETWORK:-}
|
|
|
|
# With the openvswitch plugin, if using VLANs for tenant networks,
|
|
# or if using flat or VLAN provider networks, set in localrc to
|
|
# the name of the OVS bridge to use for the physical network. The
|
|
# bridge will be created if it does not already exist, but a
|
|
# physical interface must be manually added to the bridge as a
|
|
# port for external connectivity.
|
|
#
|
|
# Example: ``OVS_PHYSICAL_BRIDGE=br-eth1``
|
|
OVS_PHYSICAL_BRIDGE=${OVS_PHYSICAL_BRIDGE:-}
|
|
|
|
# With the linuxbridge plugin, if using VLANs for tenant networks,
|
|
# or if using flat or VLAN provider networks, set in localrc to
|
|
# the name of the network interface to use for the physical
|
|
# network.
|
|
#
|
|
# Example: ``LB_PHYSICAL_INTERFACE=eth1``
|
|
LB_PHYSICAL_INTERFACE=${LB_PHYSICAL_INTERFACE:-}
|
|
|
|
# With the openvswitch plugin, set to True in localrc to enable
|
|
# provider GRE tunnels when ``ENABLE_TENANT_TUNNELS`` is False.
|
|
#
|
|
# Example: ``OVS_ENABLE_TUNNELING=True``
|
|
OVS_ENABLE_TUNNELING=${OVS_ENABLE_TUNNELING:-$ENABLE_TENANT_TUNNELS}
|
|
|
|
# Put config files in ``/etc/quantum`` for everyone to find
|
|
if [[ ! -d /etc/quantum ]]; then
|
|
sudo mkdir -p /etc/quantum
|
|
fi
|
|
sudo chown `whoami` /etc/quantum
|
|
|
|
if [[ "$Q_PLUGIN" = "openvswitch" ]]; then
|
|
Q_PLUGIN_CONF_PATH=etc/quantum/plugins/openvswitch
|
|
Q_PLUGIN_CONF_FILENAME=ovs_quantum_plugin.ini
|
|
Q_DB_NAME="ovs_quantum"
|
|
Q_PLUGIN_CLASS="quantum.plugins.openvswitch.ovs_quantum_plugin.OVSQuantumPluginV2"
|
|
elif [[ "$Q_PLUGIN" = "linuxbridge" ]]; then
|
|
Q_PLUGIN_CONF_PATH=etc/quantum/plugins/linuxbridge
|
|
Q_PLUGIN_CONF_FILENAME=linuxbridge_conf.ini
|
|
Q_DB_NAME="quantum_linux_bridge"
|
|
Q_PLUGIN_CLASS="quantum.plugins.linuxbridge.lb_quantum_plugin.LinuxBridgePluginV2"
|
|
elif [[ "$Q_PLUGIN" = "ryu" ]]; then
|
|
Q_PLUGIN_CONF_PATH=etc/quantum/plugins/ryu
|
|
Q_PLUGIN_CONF_FILENAME=ryu.ini
|
|
Q_DB_NAME="ovs_quantum"
|
|
Q_PLUGIN_CLASS="quantum.plugins.ryu.ryu_quantum_plugin.RyuQuantumPluginV2"
|
|
fi
|
|
|
|
if [[ $Q_PLUGIN_CONF_PATH == '' || $Q_PLUGIN_CONF_FILENAME == '' || $Q_PLUGIN_CLASS == '' ]]; then
|
|
echo "Quantum plugin not set.. exiting"
|
|
exit 1
|
|
fi
|
|
|
|
# If needed, move config file from ``$QUANTUM_DIR/etc/quantum`` to ``/etc/quantum``
|
|
mkdir -p /$Q_PLUGIN_CONF_PATH
|
|
Q_PLUGIN_CONF_FILE=$Q_PLUGIN_CONF_PATH/$Q_PLUGIN_CONF_FILENAME
|
|
cp $QUANTUM_DIR/$Q_PLUGIN_CONF_FILE /$Q_PLUGIN_CONF_FILE
|
|
|
|
database_connection_url dburl $Q_DB_NAME
|
|
iniset /$Q_PLUGIN_CONF_FILE DATABASE sql_connection $dburl
|
|
unset dburl
|
|
|
|
Q_CONF_FILE=/etc/quantum/quantum.conf
|
|
cp $QUANTUM_DIR/etc/quantum.conf $Q_CONF_FILE
|
|
Q_RR_CONF_FILE=/etc/quantum/rootwrap.conf
|
|
if [[ "$Q_USE_ROOTWRAP" == "False" ]]; then
|
|
Q_RR_COMMAND="sudo"
|
|
else
|
|
Q_RR_COMMAND="sudo $QUANTUM_DIR/bin/quantum-rootwrap $Q_RR_CONF_FILE"
|
|
fi
|
|
cp -p $QUANTUM_DIR/etc/rootwrap.conf $Q_RR_CONF_FILE
|
|
|
|
# Copy over the config and filter bits
|
|
Q_CONF_ROOTWRAP_D=/etc/quantum/rootwrap.d
|
|
mkdir -p $Q_CONF_ROOTWRAP_D
|
|
cp -pr $QUANTUM_DIR/etc/quantum/rootwrap.d/* $Q_CONF_ROOTWRAP_D/
|
|
fi
|
|
|
|
# Quantum service (for controller node)
|
|
if is_service_enabled q-svc; then
|
|
Q_API_PASTE_FILE=/etc/quantum/api-paste.ini
|
|
Q_POLICY_FILE=/etc/quantum/policy.json
|
|
|
|
cp $QUANTUM_DIR/etc/api-paste.ini $Q_API_PASTE_FILE
|
|
cp $QUANTUM_DIR/etc/policy.json $Q_POLICY_FILE
|
|
|
|
if is_service_enabled $DATABASE_BACKENDS; then
|
|
recreate_database $Q_DB_NAME utf8
|
|
else
|
|
echo "A database must be enabled in order to use the $Q_PLUGIN Quantum plugin."
|
|
exit 1
|
|
fi
|
|
|
|
# Update either configuration file with plugin
|
|
iniset $Q_CONF_FILE DEFAULT core_plugin $Q_PLUGIN_CLASS
|
|
|
|
iniset $Q_CONF_FILE DEFAULT auth_strategy $Q_AUTH_STRATEGY
|
|
quantum_setup_keystone $Q_API_PASTE_FILE filter:authtoken
|
|
|
|
# Configure plugin
|
|
if [[ "$Q_PLUGIN" = "openvswitch" ]]; then
|
|
if [[ "$ENABLE_TENANT_TUNNELS" = "True" ]]; then
|
|
iniset /$Q_PLUGIN_CONF_FILE OVS tenant_network_type gre
|
|
iniset /$Q_PLUGIN_CONF_FILE OVS tunnel_id_ranges $TENANT_TUNNEL_RANGES
|
|
elif [[ "$ENABLE_TENANT_VLANS" = "True" ]]; then
|
|
iniset /$Q_PLUGIN_CONF_FILE OVS tenant_network_type vlan
|
|
else
|
|
echo "WARNING - The openvswitch plugin is using local tenant networks, with no connectivity between hosts."
|
|
fi
|
|
|
|
# Override ``OVS_VLAN_RANGES`` and ``OVS_BRIDGE_MAPPINGS`` in ``localrc``
|
|
# for more complex physical network configurations.
|
|
if [[ "$OVS_VLAN_RANGES" = "" ]] && [[ "$PHYSICAL_NETWORK" != "" ]]; then
|
|
OVS_VLAN_RANGES=$PHYSICAL_NETWORK
|
|
if [[ "$TENANT_VLAN_RANGE" != "" ]]; then
|
|
OVS_VLAN_RANGES=$OVS_VLAN_RANGES:$TENANT_VLAN_RANGE
|
|
fi
|
|
fi
|
|
if [[ "$OVS_VLAN_RANGES" != "" ]]; then
|
|
iniset /$Q_PLUGIN_CONF_FILE OVS network_vlan_ranges $OVS_VLAN_RANGES
|
|
fi
|
|
|
|
# Enable tunnel networks if selected
|
|
if [[ $OVS_ENABLE_TUNNELING = "True" ]]; then
|
|
iniset /$Q_PLUGIN_CONF_FILE OVS enable_tunneling True
|
|
fi
|
|
elif [[ "$Q_PLUGIN" = "linuxbridge" ]]; then
|
|
if [[ "$ENABLE_TENANT_VLANS" = "True" ]]; then
|
|
iniset /$Q_PLUGIN_CONF_FILE VLANS tenant_network_type vlan
|
|
else
|
|
echo "WARNING - The linuxbridge plugin is using local tenant networks, with no connectivity between hosts."
|
|
fi
|
|
|
|
# Override ``LB_VLAN_RANGES`` and ``LB_INTERFACE_MAPPINGS`` in ``localrc``
|
|
# for more complex physical network configurations.
|
|
if [[ "$LB_VLAN_RANGES" = "" ]] && [[ "$PHYSICAL_NETWORK" != "" ]]; then
|
|
LB_VLAN_RANGES=$PHYSICAL_NETWORK
|
|
if [[ "$TENANT_VLAN_RANGE" != "" ]]; then
|
|
LB_VLAN_RANGES=$LB_VLAN_RANGES:$TENANT_VLAN_RANGE
|
|
fi
|
|
fi
|
|
if [[ "$LB_VLAN_RANGES" != "" ]]; then
|
|
iniset /$Q_PLUGIN_CONF_FILE VLANS network_vlan_ranges $LB_VLAN_RANGES
|
|
fi
|
|
elif [[ "$Q_PLUGIN" = "ryu" ]]; then
|
|
iniset /$Q_PLUGIN_CONF_FILE OVS openflow_controller $RYU_OFP_HOST:$RYU_OFP_PORT
|
|
iniset /$Q_PLUGIN_CONF_FILE OVS openflow_rest_api $RYU_API_HOST:$RYU_API_PORT
|
|
fi
|
|
fi
|
|
|
|
# Quantum agent (for compute nodes)
|
|
if is_service_enabled q-agt; then
|
|
# Configure agent for plugin
|
|
if [[ "$Q_PLUGIN" = "openvswitch" ]]; then
|
|
# Setup integration bridge
|
|
OVS_BRIDGE=${OVS_BRIDGE:-br-int}
|
|
quantum_setup_ovs_bridge $OVS_BRIDGE
|
|
|
|
# Setup agent for tunneling
|
|
if [[ "$OVS_ENABLE_TUNNELING" = "True" ]]; then
|
|
# Verify tunnels are supported
|
|
# REVISIT - also check kernel module support for GRE and patch ports
|
|
OVS_VERSION=`ovs-vsctl --version | head -n 1 | awk '{print $4;}'`
|
|
if [ $OVS_VERSION \< "1.4" ] && ! is_service_enabled q-svc ; then
|
|
echo "You are running OVS version $OVS_VERSION."
|
|
echo "OVS 1.4+ is required for tunneling between multiple hosts."
|
|
exit 1
|
|
fi
|
|
iniset /$Q_PLUGIN_CONF_FILE OVS enable_tunneling True
|
|
iniset /$Q_PLUGIN_CONF_FILE OVS local_ip $HOST_IP
|
|
fi
|
|
|
|
# Setup physical network bridge mappings. Override
|
|
# ``OVS_VLAN_RANGES`` and ``OVS_BRIDGE_MAPPINGS`` in ``localrc`` for more
|
|
# complex physical network configurations.
|
|
if [[ "$OVS_BRIDGE_MAPPINGS" = "" ]] && [[ "$PHYSICAL_NETWORK" != "" ]] && [[ "$OVS_PHYSICAL_BRIDGE" != "" ]]; then
|
|
OVS_BRIDGE_MAPPINGS=$PHYSICAL_NETWORK:$OVS_PHYSICAL_BRIDGE
|
|
|
|
# Configure bridge manually with physical interface as port for multi-node
|
|
sudo ovs-vsctl --no-wait -- --may-exist add-br $OVS_PHYSICAL_BRIDGE
|
|
fi
|
|
if [[ "$OVS_BRIDGE_MAPPINGS" != "" ]]; then
|
|
iniset /$Q_PLUGIN_CONF_FILE OVS bridge_mappings $OVS_BRIDGE_MAPPINGS
|
|
fi
|
|
AGENT_BINARY="$QUANTUM_DIR/bin/quantum-openvswitch-agent"
|
|
elif [[ "$Q_PLUGIN" = "linuxbridge" ]]; then
|
|
# Setup physical network interface mappings. Override
|
|
# ``LB_VLAN_RANGES`` and ``LB_INTERFACE_MAPPINGS`` in ``localrc`` for more
|
|
# complex physical network configurations.
|
|
if [[ "$LB_INTERFACE_MAPPINGS" = "" ]] && [[ "$PHYSICAL_NETWORK" != "" ]] && [[ "$LB_PHYSICAL_INTERFACE" != "" ]]; then
|
|
LB_INTERFACE_MAPPINGS=$PHYSICAL_NETWORK:$LB_PHYSICAL_INTERFACE
|
|
fi
|
|
if [[ "$LB_INTERFACE_MAPPINGS" != "" ]]; then
|
|
iniset /$Q_PLUGIN_CONF_FILE LINUX_BRIDGE physical_interface_mappings $LB_INTERFACE_MAPPINGS
|
|
fi
|
|
AGENT_BINARY="$QUANTUM_DIR/bin/quantum-linuxbridge-agent"
|
|
elif [[ "$Q_PLUGIN" = "ryu" ]]; then
|
|
# Set up integration bridge
|
|
OVS_BRIDGE=${OVS_BRIDGE:-br-int}
|
|
quantum_setup_ovs_bridge $OVS_BRIDGE
|
|
if [ -n "$RYU_INTERNAL_INTERFACE" ]; then
|
|
sudo ovs-vsctl --no-wait -- --may-exist add-port $OVS_BRIDGE $RYU_INTERNAL_INTERFACE
|
|
fi
|
|
AGENT_BINARY="$QUANTUM_DIR/quantum/plugins/ryu/agent/ryu_quantum_agent.py"
|
|
fi
|
|
# Update config w/rootwrap
|
|
iniset /$Q_PLUGIN_CONF_FILE AGENT root_helper "$Q_RR_COMMAND"
|
|
fi
|
|
|
|
# Quantum DHCP
|
|
if is_service_enabled q-dhcp; then
|
|
AGENT_DHCP_BINARY="$QUANTUM_DIR/bin/quantum-dhcp-agent"
|
|
|
|
Q_DHCP_CONF_FILE=/etc/quantum/dhcp_agent.ini
|
|
|
|
cp $QUANTUM_DIR/etc/dhcp_agent.ini $Q_DHCP_CONF_FILE
|
|
|
|
# Set verbose
|
|
iniset $Q_DHCP_CONF_FILE DEFAULT verbose True
|
|
# Set debug
|
|
iniset $Q_DHCP_CONF_FILE DEFAULT debug True
|
|
iniset $Q_DHCP_CONF_FILE DEFAULT use_namespaces $Q_USE_NAMESPACE
|
|
iniset $Q_DHCP_CONF_FILE DEFAULT state_path $DATA_DIR/quantum
|
|
|
|
quantum_setup_keystone $Q_DHCP_CONF_FILE DEFAULT set_auth_url
|
|
|
|
# Update config w/rootwrap
|
|
iniset $Q_DHCP_CONF_FILE DEFAULT root_helper "$Q_RR_COMMAND"
|
|
|
|
if [[ "$Q_PLUGIN" = "openvswitch" ]]; then
|
|
iniset $Q_DHCP_CONF_FILE DEFAULT interface_driver quantum.agent.linux.interface.OVSInterfaceDriver
|
|
elif [[ "$Q_PLUGIN" = "linuxbridge" ]]; then
|
|
iniset $Q_DHCP_CONF_FILE DEFAULT interface_driver quantum.agent.linux.interface.BridgeInterfaceDriver
|
|
elif [[ "$Q_PLUGIN" = "ryu" ]]; then
|
|
iniset $Q_DHCP_CONF_FILE DEFAULT interface_driver quantum.agent.linux.interface.RyuInterfaceDriver
|
|
iniset $Q_DHCP_CONF_FILE DEFAULT ryu_api_host $RYU_API_HOST:$RYU_API_PORT
|
|
fi
|
|
fi
|
|
|
|
# Quantum L3
|
|
if is_service_enabled q-l3; then
|
|
AGENT_L3_BINARY="$QUANTUM_DIR/bin/quantum-l3-agent"
|
|
PUBLIC_BRIDGE=${PUBLIC_BRIDGE:-br-ex}
|
|
Q_L3_CONF_FILE=/etc/quantum/l3_agent.ini
|
|
|
|
cp $QUANTUM_DIR/etc/l3_agent.ini $Q_L3_CONF_FILE
|
|
|
|
# Set verbose
|
|
iniset $Q_L3_CONF_FILE DEFAULT verbose True
|
|
# Set debug
|
|
iniset $Q_L3_CONF_FILE DEFAULT debug True
|
|
|
|
iniset $Q_L3_CONF_FILE DEFAULT metadata_ip $Q_META_DATA_IP
|
|
iniset $Q_L3_CONF_FILE DEFAULT use_namespaces $Q_USE_NAMESPACE
|
|
|
|
iniset $Q_L3_CONF_FILE DEFAULT root_helper "$Q_RR_COMMAND"
|
|
|
|
quantum_setup_keystone $Q_L3_CONF_FILE DEFAULT set_auth_url
|
|
if [[ "$Q_PLUGIN" == "openvswitch" ]]; then
|
|
iniset $Q_L3_CONF_FILE DEFAULT interface_driver quantum.agent.linux.interface.OVSInterfaceDriver
|
|
iniset $Q_L3_CONF_FILE DEFAULT external_network_bridge $PUBLIC_BRIDGE
|
|
# Set up external bridge
|
|
quantum_setup_external_bridge $PUBLIC_BRIDGE
|
|
elif [[ "$Q_PLUGIN" = "linuxbridge" ]]; then
|
|
iniset $Q_L3_CONF_FILE DEFAULT interface_driver quantum.agent.linux.interface.BridgeInterfaceDriver
|
|
iniset $Q_L3_CONF_FILE DEFAULT external_network_bridge ''
|
|
elif [[ "$Q_PLUGIN" = "ryu" ]]; then
|
|
iniset $Q_L3_CONF_FILE DEFAULT interface_driver quantum.agent.linux.interface.RyuInterfaceDriver
|
|
iniset $Q_L3_CONF_FILE DEFAULT external_network_bridge $PUBLIC_BRIDGE
|
|
iniset $Q_L3_CONF_FILE DEFAULT ryu_api_host $RYU_API_HOST:$RYU_API_PORT
|
|
# Set up external bridge
|
|
quantum_setup_external_bridge $PUBLIC_BRIDGE
|
|
fi
|
|
fi
|
|
|
|
# Quantum RPC support - must be updated prior to starting any of the services
|
|
if is_service_enabled quantum; then
|
|
iniset $Q_CONF_FILE DEFAULT control_exchange quantum
|
|
if is_service_enabled qpid ; then
|
|
iniset $Q_CONF_FILE DEFAULT rpc_backend quantum.openstack.common.rpc.impl_qpid
|
|
elif is_service_enabled zeromq; then
|
|
iniset $Q_CONF_FILE DEFAULT rpc_backend quantum.openstack.common.rpc.impl_zmq
|
|
elif [ -n "$RABBIT_HOST" ] && [ -n "$RABBIT_PASSWORD" ]; then
|
|
iniset $Q_CONF_FILE DEFAULT rabbit_host $RABBIT_HOST
|
|
iniset $Q_CONF_FILE DEFAULT rabbit_password $RABBIT_PASSWORD
|
|
fi
|
|
if [[ "$Q_USE_DEBUG_COMMAND" == "True" ]]; then
|
|
Q_DEBUG_CONF_FILE=/etc/quantum/debug.ini
|
|
cp $QUANTUM_DIR/etc/l3_agent.ini $Q_DEBUG_CONF_FILE
|
|
iniset $Q_L3_CONF_FILE DEFAULT verbose False
|
|
iniset $Q_L3_CONF_FILE DEFAULT debug False
|
|
iniset $Q_L3_CONF_FILE DEFAULT metadata_ip $Q_META_DATA_IP
|
|
iniset $Q_L3_CONF_FILE DEFAULT use_namespaces $Q_USE_NAMESPACE
|
|
iniset $Q_L3_CONF_FILE DEFAULT root_helper "sudo"
|
|
fi
|
|
fi
|
|
|
|
|
|
# Nova
|
|
# ----
|
|
|
|
if is_service_enabled nova; then
|
|
echo_summary "Configuring Nova"
|
|
configure_nova
|
|
fi
|
|
|
|
if is_service_enabled n-net q-dhcp; then
|
|
# Delete traces of nova networks from prior runs
|
|
sudo killall dnsmasq || true
|
|
clean_iptables
|
|
rm -rf $NOVA_STATE_PATH/networks
|
|
mkdir -p $NOVA_STATE_PATH/networks
|
|
|
|
# Force IP forwarding on, just on case
|
|
sudo sysctl -w net.ipv4.ip_forward=1
|
|
fi
|
|
|
|
|
|
# Storage Service
|
|
# ---------------
|
|
|
|
if is_service_enabled swift; then
|
|
echo_summary "Configuring Swift"
|
|
init_swift
|
|
fi
|
|
|
|
|
|
# Volume Service
|
|
# --------------
|
|
|
|
if is_service_enabled cinder; then
|
|
echo_summary "Configuring Cinder"
|
|
init_cinder
|
|
fi
|
|
|
|
if is_service_enabled nova; then
|
|
echo_summary "Configuring Nova"
|
|
# Rebuild the config file from scratch
|
|
create_nova_conf
|
|
init_nova
|
|
|
|
# Additional Nova configuration that is dependent on other services
|
|
if is_service_enabled quantum; then
|
|
add_nova_opt "network_api_class=nova.network.quantumv2.api.API"
|
|
add_nova_opt "quantum_admin_username=$Q_ADMIN_USERNAME"
|
|
add_nova_opt "quantum_admin_password=$SERVICE_PASSWORD"
|
|
add_nova_opt "quantum_admin_auth_url=$KEYSTONE_SERVICE_PROTOCOL://$KEYSTONE_SERVICE_HOST:$KEYSTONE_AUTH_PORT/v2.0"
|
|
add_nova_opt "quantum_auth_strategy=$Q_AUTH_STRATEGY"
|
|
add_nova_opt "quantum_admin_tenant_name=$SERVICE_TENANT_NAME"
|
|
add_nova_opt "quantum_url=http://$Q_HOST:$Q_PORT"
|
|
|
|
if [[ "$Q_PLUGIN" = "openvswitch" ]]; then
|
|
NOVA_VIF_DRIVER=${NOVA_VIF_DRIVER:-"nova.virt.libvirt.vif.LibvirtHybridOVSBridgeDriver"}
|
|
elif [[ "$Q_PLUGIN" = "linuxbridge" ]]; then
|
|
NOVA_VIF_DRIVER=${NOVA_VIF_DRIVER:-"nova.virt.libvirt.vif.QuantumLinuxBridgeVIFDriver"}
|
|
elif [[ "$Q_PLUGIN" = "ryu" ]]; then
|
|
NOVA_VIF_DRIVER=${NOVA_VIF_DRIVER:-"quantum.plugins.ryu.nova.vif.LibvirtOpenVswitchOFPRyuDriver"}
|
|
add_nova_opt "libvirt_ovs_integration_bridge=$OVS_BRIDGE"
|
|
add_nova_opt "linuxnet_ovs_ryu_api_host=$RYU_API_HOST:$RYU_API_PORT"
|
|
add_nova_opt "libvirt_ovs_ryu_api_host=$RYU_API_HOST:$RYU_API_PORT"
|
|
fi
|
|
add_nova_opt "libvirt_vif_driver=$NOVA_VIF_DRIVER"
|
|
add_nova_opt "linuxnet_interface_driver=$LINUXNET_VIF_DRIVER"
|
|
elif is_service_enabled n-net; then
|
|
add_nova_opt "network_manager=nova.network.manager.$NET_MAN"
|
|
add_nova_opt "public_interface=$PUBLIC_INTERFACE"
|
|
add_nova_opt "vlan_interface=$VLAN_INTERFACE"
|
|
add_nova_opt "flat_network_bridge=$FLAT_NETWORK_BRIDGE"
|
|
if [ -n "$FLAT_INTERFACE" ]; then
|
|
add_nova_opt "flat_interface=$FLAT_INTERFACE"
|
|
fi
|
|
fi
|
|
# All nova-compute workers need to know the vnc configuration options
|
|
# These settings don't hurt anything if n-xvnc and n-novnc are disabled
|
|
if is_service_enabled n-cpu; then
|
|
NOVNCPROXY_URL=${NOVNCPROXY_URL:-"http://$SERVICE_HOST:6080/vnc_auto.html"}
|
|
add_nova_opt "novncproxy_base_url=$NOVNCPROXY_URL"
|
|
XVPVNCPROXY_URL=${XVPVNCPROXY_URL:-"http://$SERVICE_HOST:6081/console"}
|
|
add_nova_opt "xvpvncproxy_base_url=$XVPVNCPROXY_URL"
|
|
fi
|
|
if [ "$VIRT_DRIVER" = 'xenserver' ]; then
|
|
VNCSERVER_PROXYCLIENT_ADDRESS=${VNCSERVER_PROXYCLIENT_ADDRESS=169.254.0.1}
|
|
else
|
|
VNCSERVER_PROXYCLIENT_ADDRESS=${VNCSERVER_PROXYCLIENT_ADDRESS=127.0.0.1}
|
|
fi
|
|
# Address on which instance vncservers will listen on compute hosts.
|
|
# For multi-host, this should be the management ip of the compute host.
|
|
VNCSERVER_LISTEN=${VNCSERVER_LISTEN=127.0.0.1}
|
|
add_nova_opt "vncserver_listen=$VNCSERVER_LISTEN"
|
|
add_nova_opt "vncserver_proxyclient_address=$VNCSERVER_PROXYCLIENT_ADDRESS"
|
|
add_nova_opt "ec2_dmz_host=$EC2_DMZ_HOST"
|
|
if is_service_enabled zeromq; then
|
|
add_nova_opt "rpc_backend=nova.openstack.common.rpc.impl_zmq"
|
|
elif is_service_enabled qpid; then
|
|
add_nova_opt "rpc_backend=nova.rpc.impl_qpid"
|
|
elif [ -n "$RABBIT_HOST" ] && [ -n "$RABBIT_PASSWORD" ]; then
|
|
add_nova_opt "rabbit_host=$RABBIT_HOST"
|
|
add_nova_opt "rabbit_password=$RABBIT_PASSWORD"
|
|
fi
|
|
add_nova_opt "glance_api_servers=$GLANCE_HOSTPORT"
|
|
|
|
|
|
# XenServer
|
|
# ---------
|
|
|
|
if [ "$VIRT_DRIVER" = 'xenserver' ]; then
|
|
echo_summary "Using XenServer virtualization driver"
|
|
read_password XENAPI_PASSWORD "ENTER A PASSWORD TO USE FOR XEN."
|
|
add_nova_opt "compute_driver=xenapi.XenAPIDriver"
|
|
XENAPI_CONNECTION_URL=${XENAPI_CONNECTION_URL:-"http://169.254.0.1"}
|
|
XENAPI_USER=${XENAPI_USER:-"root"}
|
|
add_nova_opt "xenapi_connection_url=$XENAPI_CONNECTION_URL"
|
|
add_nova_opt "xenapi_connection_username=$XENAPI_USER"
|
|
add_nova_opt "xenapi_connection_password=$XENAPI_PASSWORD"
|
|
add_nova_opt "flat_injected=False"
|
|
# Need to avoid crash due to new firewall support
|
|
XEN_FIREWALL_DRIVER=${XEN_FIREWALL_DRIVER:-"nova.virt.firewall.IptablesFirewallDriver"}
|
|
add_nova_opt "firewall_driver=$XEN_FIREWALL_DRIVER"
|
|
elif [ "$VIRT_DRIVER" = 'openvz' ]; then
|
|
echo_summary "Using OpenVZ virtualization driver"
|
|
# TODO(deva): OpenVZ driver does not yet work if compute_driver is set here.
|
|
# Replace connection_type when this is fixed.
|
|
# add_nova_opt "compute_driver=openvz.connection.OpenVzConnection"
|
|
add_nova_opt "connection_type=openvz"
|
|
LIBVIRT_FIREWALL_DRIVER=${LIBVIRT_FIREWALL_DRIVER:-"nova.virt.libvirt.firewall.IptablesFirewallDriver"}
|
|
add_nova_opt "firewall_driver=$LIBVIRT_FIREWALL_DRIVER"
|
|
else
|
|
echo_summary "Using libvirt virtualization driver"
|
|
add_nova_opt "compute_driver=libvirt.LibvirtDriver"
|
|
LIBVIRT_FIREWALL_DRIVER=${LIBVIRT_FIREWALL_DRIVER:-"nova.virt.libvirt.firewall.IptablesFirewallDriver"}
|
|
add_nova_opt "firewall_driver=$LIBVIRT_FIREWALL_DRIVER"
|
|
fi
|
|
fi
|
|
|
|
|
|
# Launch Services
|
|
# ===============
|
|
|
|
# Only run the services specified in ``ENABLED_SERVICES``
|
|
|
|
# Launch Swift Services
|
|
if is_service_enabled swift; then
|
|
echo_summary "Starting Swift"
|
|
start_swift
|
|
fi
|
|
|
|
# Launch the Glance services
|
|
if is_service_enabled g-api g-reg; then
|
|
echo_summary "Starting Glance"
|
|
start_glance
|
|
fi
|
|
|
|
# Create an access key and secret key for nova ec2 register image
|
|
if is_service_enabled key && is_service_enabled swift3 && is_service_enabled nova; then
|
|
NOVA_USER_ID=$(keystone user-list | grep ' nova ' | get_field 1)
|
|
NOVA_TENANT_ID=$(keystone tenant-list | grep " $SERVICE_TENANT_NAME " | get_field 1)
|
|
CREDS=$(keystone ec2-credentials-create --user_id $NOVA_USER_ID --tenant_id $NOVA_TENANT_ID)
|
|
ACCESS_KEY=$(echo "$CREDS" | awk '/ access / { print $4 }')
|
|
SECRET_KEY=$(echo "$CREDS" | awk '/ secret / { print $4 }')
|
|
add_nova_opt "s3_access_key=$ACCESS_KEY"
|
|
add_nova_opt "s3_secret_key=$SECRET_KEY"
|
|
add_nova_opt "s3_affix_tenant=True"
|
|
fi
|
|
|
|
screen_it zeromq "cd $NOVA_DIR && $NOVA_DIR/bin/nova-rpc-zmq-receiver"
|
|
|
|
# Launch the nova-api and wait for it to answer before continuing
|
|
if is_service_enabled n-api; then
|
|
echo_summary "Starting Nova API"
|
|
screen_it n-api "cd $NOVA_DIR && $NOVA_BIN_DIR/nova-api"
|
|
echo "Waiting for nova-api to start..."
|
|
if ! timeout $SERVICE_TIMEOUT sh -c "while ! http_proxy= wget -q -O- http://127.0.0.1:8774; do sleep 1; done"; then
|
|
echo "nova-api did not start"
|
|
exit 1
|
|
fi
|
|
fi
|
|
|
|
if is_service_enabled q-svc; then
|
|
echo_summary "Starting Quantum"
|
|
# Start the Quantum service
|
|
screen_it q-svc "cd $QUANTUM_DIR && python $QUANTUM_DIR/bin/quantum-server --config-file $Q_CONF_FILE --config-file /$Q_PLUGIN_CONF_FILE"
|
|
echo "Waiting for Quantum to start..."
|
|
if ! timeout $SERVICE_TIMEOUT sh -c "while ! http_proxy= wget -q -O- http://127.0.0.1:9696; do sleep 1; done"; then
|
|
echo "Quantum did not start"
|
|
exit 1
|
|
fi
|
|
|
|
# Configure Quantum elements
|
|
# Configure internal network & subnet
|
|
|
|
TENANT_ID=$(keystone tenant-list | grep " demo " | get_field 1)
|
|
|
|
# Create a small network
|
|
# Since quantum command is executed in admin context at this point,
|
|
# ``--tenant_id`` needs to be specified.
|
|
NET_ID=$(quantum net-create --tenant_id $TENANT_ID "$PRIVATE_NETWORK_NAME" | grep ' id ' | get_field 2)
|
|
SUBNET_ID=$(quantum subnet-create --tenant_id $TENANT_ID --ip_version 4 --gateway $NETWORK_GATEWAY $NET_ID $FIXED_RANGE | grep ' id ' | get_field 2)
|
|
if is_service_enabled q-l3; then
|
|
# Create a router, and add the private subnet as one of its interfaces
|
|
ROUTER_ID=$(quantum router-create --tenant_id $TENANT_ID router1 | grep ' id ' | get_field 2)
|
|
quantum router-interface-add $ROUTER_ID $SUBNET_ID
|
|
# Create an external network, and a subnet. Configure the external network as router gw
|
|
EXT_NET_ID=$(quantum net-create "$PUBLIC_NETWORK_NAME" -- --router:external=True | grep ' id ' | get_field 2)
|
|
EXT_GW_IP=$(quantum subnet-create --ip_version 4 $EXT_NET_ID $FLOATING_RANGE -- --enable_dhcp=False | grep 'gateway_ip' | get_field 2)
|
|
quantum router-gateway-set $ROUTER_ID $EXT_NET_ID
|
|
if is_quantum_ovs_base_plugin "$Q_PLUGIN" && [[ "$Q_USE_NAMESPACE" = "True" ]]; then
|
|
CIDR_LEN=${FLOATING_RANGE#*/}
|
|
sudo ip addr add $EXT_GW_IP/$CIDR_LEN dev $PUBLIC_BRIDGE
|
|
sudo ip link set $PUBLIC_BRIDGE up
|
|
ROUTER_GW_IP=`quantum port-list -c fixed_ips -c device_owner | grep router_gateway | awk -F '"' '{ print $8; }'`
|
|
sudo route add -net $FIXED_RANGE gw $ROUTER_GW_IP
|
|
fi
|
|
if [[ "$Q_USE_NAMESPACE" == "False" ]]; then
|
|
# Explicitly set router id in l3 agent configuration
|
|
iniset $Q_L3_CONF_FILE DEFAULT router_id $ROUTER_ID
|
|
fi
|
|
fi
|
|
|
|
elif is_service_enabled $DATABASE_BACKENDS && is_service_enabled n-net; then
|
|
# Create a small network
|
|
$NOVA_BIN_DIR/nova-manage network create "$PRIVATE_NETWORK_NAME" $FIXED_RANGE 1 $FIXED_NETWORK_SIZE $NETWORK_CREATE_ARGS
|
|
|
|
# Create some floating ips
|
|
$NOVA_BIN_DIR/nova-manage floating create $FLOATING_RANGE --pool=$PUBLIC_NETWORK
|
|
|
|
# Create a second pool
|
|
$NOVA_BIN_DIR/nova-manage floating create --ip_range=$TEST_FLOATING_RANGE --pool=$TEST_FLOATING_POOL
|
|
fi
|
|
|
|
# Start up the quantum agents if enabled
|
|
screen_it q-agt "python $AGENT_BINARY --config-file $Q_CONF_FILE --config-file /$Q_PLUGIN_CONF_FILE"
|
|
screen_it q-dhcp "python $AGENT_DHCP_BINARY --config-file $Q_CONF_FILE --config-file=$Q_DHCP_CONF_FILE"
|
|
screen_it q-l3 "python $AGENT_L3_BINARY --config-file $Q_CONF_FILE --config-file=$Q_L3_CONF_FILE"
|
|
|
|
if is_service_enabled nova; then
|
|
echo_summary "Starting Nova"
|
|
start_nova
|
|
fi
|
|
if is_service_enabled cinder; then
|
|
echo_summary "Starting Cinder"
|
|
start_cinder
|
|
fi
|
|
if is_service_enabled ceilometer; then
|
|
echo_summary "Configuring Ceilometer"
|
|
configure_ceilometer
|
|
echo_summary "Starting Ceilometer"
|
|
start_ceilometer
|
|
fi
|
|
|
|
# Starting the nova-objectstore only if swift3 service is not enabled.
|
|
# Swift will act as s3 objectstore.
|
|
is_service_enabled swift3 || \
|
|
screen_it n-obj "cd $NOVA_DIR && $NOVA_BIN_DIR/nova-objectstore"
|
|
|
|
|
|
# Configure and launch heat engine, api and metadata
|
|
if is_service_enabled heat; then
|
|
# Initialize heat, including replacing nova flavors
|
|
echo_summary "Configuring Heat"
|
|
init_heat
|
|
echo_summary "Starting Heat"
|
|
start_heat
|
|
fi
|
|
|
|
|
|
# Install Images
|
|
# ==============
|
|
|
|
# Upload an image to glance.
|
|
#
|
|
# The default image is cirros, a small testing image which lets you login as **root**
|
|
# cirros also uses ``cloud-init``, supporting login via keypair and sending scripts as
|
|
# userdata. See https://help.ubuntu.com/community/CloudInit for more on cloud-init
|
|
#
|
|
# Override ``IMAGE_URLS`` with a comma-separated list of UEC images.
|
|
# * **oneiric**: http://uec-images.ubuntu.com/oneiric/current/oneiric-server-cloudimg-amd64.tar.gz
|
|
# * **precise**: http://uec-images.ubuntu.com/precise/current/precise-server-cloudimg-amd64.tar.gz
|
|
|
|
if is_service_enabled g-reg; then
|
|
echo_summary "Uploading images"
|
|
TOKEN=$(keystone token-get | grep ' id ' | get_field 2)
|
|
|
|
# Option to upload legacy ami-tty, which works with xenserver
|
|
if [[ -n "$UPLOAD_LEGACY_TTY" ]]; then
|
|
IMAGE_URLS="${IMAGE_URLS:+${IMAGE_URLS},}https://github.com/downloads/citrix-openstack/warehouse/tty.tgz"
|
|
fi
|
|
|
|
for image_url in ${IMAGE_URLS//,/ }; do
|
|
upload_image $image_url $TOKEN
|
|
done
|
|
fi
|
|
|
|
|
|
# Configure Tempest last to ensure that the runtime configuration of
|
|
# the various OpenStack services can be queried.
|
|
if is_service_enabled tempest; then
|
|
configure_tempest
|
|
fi
|
|
|
|
|
|
# Run local script
|
|
# ================
|
|
|
|
# Run ``local.sh`` if it exists to perform user-managed tasks
|
|
if [[ -x $TOP_DIR/local.sh ]]; then
|
|
echo "Running user script $TOP_DIR/local.sh"
|
|
$TOP_DIR/local.sh
|
|
fi
|
|
|
|
|
|
# Fin
|
|
# ===
|
|
|
|
set +o xtrace
|
|
|
|
if [[ -n "$LOGFILE" ]]; then
|
|
exec 1>&3
|
|
# Force all output to stdout and logs now
|
|
exec 1> >( tee -a "${LOGFILE}" ) 2>&1
|
|
else
|
|
# Force all output to stdout now
|
|
exec 1>&3
|
|
fi
|
|
|
|
|
|
# Using the cloud
|
|
# ---------------
|
|
|
|
echo ""
|
|
echo ""
|
|
echo ""
|
|
|
|
# If you installed Horizon on this server you should be able
|
|
# to access the site using your browser.
|
|
if is_service_enabled horizon; then
|
|
echo "Horizon is now available at http://$SERVICE_HOST/"
|
|
fi
|
|
|
|
# Warn that the default flavors have been changed by Heat
|
|
if is_service_enabled heat; then
|
|
echo "Heat has replaced the default flavors. View by running: nova flavor-list"
|
|
fi
|
|
|
|
# If Keystone is present you can point ``nova`` cli to this server
|
|
if is_service_enabled key; then
|
|
echo "Keystone is serving at $KEYSTONE_AUTH_PROTOCOL://$SERVICE_HOST:$KEYSTONE_API_PORT/v2.0/"
|
|
echo "Examples on using novaclient command line is in exercise.sh"
|
|
echo "The default users are: admin and demo"
|
|
echo "The password: $ADMIN_PASSWORD"
|
|
fi
|
|
|
|
# Echo ``HOST_IP`` - useful for ``build_uec.sh``, which uses dhcp to give the instance an address
|
|
echo "This is your host ip: $HOST_IP"
|
|
|
|
# Warn that ``EXTRA_FLAGS`` needs to be converted to ``EXTRA_OPTS``
|
|
if [[ -n "$EXTRA_FLAGS" ]]; then
|
|
echo_summary "WARNING: EXTRA_FLAGS is defined and may need to be converted to EXTRA_OPTS"
|
|
fi
|
|
|
|
# Indicate how long this took to run (bash maintained variable ``SECONDS``)
|
|
echo_summary "stack.sh completed in $SECONDS seconds."
|