Ensure all references to the Ansible control host are worded as such, to ensure consistency and avoid potential confusion with the OpenStack controllers. Change-Id: Id92e537ccbfdd55287b8eae296f649640c70ce17
11 KiB
Release Notes
Release notes are now built using reno, and will be available separately from the kayobe documentation for future releases.
Kayobe 3.1.0
Kayobe 3.1.0 was released on 22nd February 2018 and is based on the Pike release of OpenStack.
Features
- Adds
--interface-limit
and--interface-description-limit
arguments to thekayobe physical network configure
command. These arguments allow configuration to be limited to a subset of switch interfaces. - Adds a
--display
argument tokayobe physical network configure
command. This will output the candidate switch configuration without applying it. - Adds support for configuration of custom fluentd filters, and
additional config file templates for heat, ironic, keystone, magnum,
murano, sahara, and swift in
$KAYOBE_CONFIG_PATH/kolla/config/<component>/
. - Adds support for specifying a local Yum mirror for package installation.
- Adds the command
kayobe network connectivity check
which can be used to verify network connectivity in the cloud hosts. - Adds a variable
kolla_nova_compute_ironic_host
which may be used to set which hosts run the nova compute service for ironic. This may be used to avoid the experimental HA nova compute service for ironic, by specifying a single host. - Adds support for deployment of virtualised compute hosts. These
hosts should be added to the
[compute]
group. - Adds support for multiple external networks.
external_net_names
should be a list of names of networks. - Adds commands for management of baremetal compute nodes -
kayobe baremetal compute inspect
,kayobe baremetal compute manage
, andkayobe baremetal compute provide
. - Adds support for installation and use of a python virtual
environment for remote execution of ansible modules, providing isolation
from the system's python packages. This is enabled by setting a host
variable,
ansible_python_interpreter
, to the path to a python interpreter in a virtualenv, noting that Jinja2 templating is not supported for this variable. - Adds support for configuration of a python virtual environment for
remote execution of ansible modules in kolla-ansible, providing
isolation from the system's python packages. This is enabled by setting
the variable
kolla_ansible_target_venv
to a path to the virtualenv. The default for this variable is{{ virtualenv_path }}/kolla-ansible
. - Adds tags to plays to support more fine grained configuration using
the
--tags
argument. - Adds support for deployment of storage hosts. These hosts should be
added to the
[storage]
group. - Adds support for the tagging of ceph disks.
- Adds support for post-deployment configuration of Grafana data sources and dashboards.
Upgrade Notes
Modifies the default value for
inspector_manage_firewall
fromFalse
toTrue
. Management of the firewall by ironic inspector is important to ensure that DHCP offers are not made to nodes during provisioning by inspector's DHCP server.Disables swift by default. The default value of
kolla_enable_swift
is nowno
.The default list of neutron ML2 mechanism drivers (
kolla_neutron_ml2_mechanism_drivers
) has been removed in favour of using the defaults provided by kolla-ansible. Users relying on the default list ofopenvswitch
andgenericswitch
should set the value explicitly.Adds a variable
config_path
, used to set the base path to configuration on remote hosts. The default value is/opt/kayobe/etc
.Modifies the variable used to configure the kolla build configuration path from
kolla_config_path
tokolla_build_config_path
. This provides a cleaner separation of kolla and kolla-ansible configuration options. The default value is{{ config_path }}/kolla
.Adds a group
container-image-builders
, which defaults to containing the seed. Hosts in this group will build container images. Previously, container images for the seed were built on the seed, and container images for the overcloud were built on the controllers. The new design is intended to encourage a build, push, pull workflow.It is now possible to configure kayobe to use a virtual environment for remote execution of ansible modules. If this is required, the following commands should be run in order to ensure that the virtual environments exist on the remote hosts:
(kayobe) $ kayobe seed hypervisor host upgrade (kayobe) $ kayobe seed host upgrade (kayobe) $ kayobe overcloud host upgrade
The default behaviour is now to configure kolla-ansible to use a virtual environment for remote execution of ansible modules. In order to ensure the virtual environment exists on the remote hosts, run the following commands:
(kayobe) $ kayobe seed hypervisor host upgrade (kayobe) $ kayobe seed host upgrade (kayobe) $ kayobe overcloud host upgrade
The previous behaviour of installing python dependencies directly to the host can be used by setting
kolla_ansible_target_venv
toNone
.Adds a workaround for an issue with CentOS cloud images 7.2 (1511) onwards, which have a bogus name server entry in /etc/resolv.conf, 10.0.2.3. Cloud-init only appends name server entries to this file, and will not remove this bogus entry. Typically this leads to a delay of around 30 seconds when connecting via SSH, due to a timeout in NSS. The workaround employed here is to remove this bogus entry from the image using virt-customize, if it exists. See https://bugs.centos.org/view.php?id=14369.
Adds a group
storage
, which used for deploy node with cinder-volume, LVM or ceph-osd. If you want to add these services to compute or control group, you need to overridekolla_overcloud_inventory_storage_groups
.
Kayobe 3.0.0
Kayobe 3.0.0 was released on 20th September 2017.
Features
- Adds support for the OpenStack Pike release.
- Adds support for saving overcloud service configuration to the Ansible control host.
- Adds support for generating overcloud service configuration, without applying it to the running system.
Upgrade Notes
See the upgrade notes for the pike release of the OpenStack services in use.
Kayobe 2.0.0
Kayobe 2.0.0 was released on 15th September 2017.
Features
- Adds support for configuration of networks for out-of-band
management for the overcloud and control plane hosts via the
oob_oc_net_name
andoob_wl_net_name
variables respectively. - Adds support for configuration of a seed hypervisor host. This host runs the seed VM. Currently, configuration of host networking, NTP, and libvirt storage pools and networks is supported.
- Adds a
base_path
variable to simplify configuration of paths. This is used to set the default value ofimage_cache_path
andsource_checkout_path
. The default value of the base path may be set by the$KAYOBE_BASE_PATH
environment variable. - Adds a
virtualenv_path
variable to configure the path on which to create virtual environments. - Uses the CentOS 7 cloud image for the seed VM by default.
- Adds a command to deprovision the seed VM,
kayobe seed vm deprovision
. - Adds support for configuration of Juniper switches.
- Adds support for bonded (LAG) host network interfaces.
- Adds support for the overlay docker storage driver on the seed and overcloud hosts.
- Improves the Vagrant development environment, and provides configuration for a single controller with a single network.
- Adds support for building customised Ironic Python Agent (IPA)
deployment images using Diskimage Builder (DIB). These can be built
using the commands
kayobe seed deployment image build
andkayobe overcloud deployment image build
. - Adds a command to save overcloud introspection data,
kayobe overcloud introspection data save
. - Separates the external network into external and public networks.
The public network carries public API traffic, and is configured via
public_net_name
. - Adds a
network
group, with networking and load balancing services moved to it. The group is a subgroup of thecontrollers
group by default. - Decomposes the overcloud inventory into top level, components, and services. This allows a deployer to customise their inventory at various levels, by providing a custom inventory template for one or more sections of the inventory.
- Adds support for configuration of sysctl parameters on the seed, seed hypervisor and overcloud hosts.
- Adds an inspection-store container for storage of workload hardware inspection data in environments without Swift.
- Adds configuration of gatewys in provisioning and inspection networks.
- Adds support for free-form configuration of Glance.
- Adds support for Ubuntu Ansible control hosts.
- Adds support for passing through host variables from kayobe to
kolla-ansible. By default
ansible_host
,ansible_port
, andansible_ssh_private_key_file
.
Upgrade Notes
- It is no longer necessary to set the
seed_vm_interfaces
variable, as the seed VM's network interfaces are now determined by the standardseed_network_interfaces
variable. - If using a CentOS 7 cloud image for the seed VM, it is no longer
necessary to set the
seed_vm_root_image
variable. - The default value of
kolla_enable_haproxy
has been changed toTrue
. - If using a custom inventory, a
network
group should be added to it. If the Ansible control hosts are providing networking services, then thenetwork
group should be a subgroup of thecontrollers
group. - The
overcloud_groups
variable is now determined more intelligently, and it is generally no longer necessary to set it manually. - The provisioning network is now used to access the TFTP server during workload hardware inspection.
- A default gateway may be advertised to compute nodes during workload inspection, allowing access to an ironic inspector API on the internal API network.
Kayobe 1.1.0
Kayobe 1.1.0 was released on 17th July 2017.
Features
- Support static routes on control plane networks
- Improve documentation
- Initial support for in-development Pike release
- Upgrade kayobe Ansible control host & control plane
- Support overcloud service destroy command
- Support fluentd custom output configuration
Kayobe 1.0.0
1.0.0 is the first 'official' release of the Kayobe OpenStack deployment tool. It was released on 29th June 2017.
Features
This release includes the following features:
- Heavily automated using Ansible
kayobe
Command Line Interface (CLI) for cloud operators- Deployment of a seed VM used to manage the OpenStack control plane
- Configuration of physical network infrastructure
- Discovery, introspection and provisioning of control plane hardware using OpenStack bifrost
- Deployment of an OpenStack control plane using OpenStack kolla-ansible
- Discovery, introspection and provisioning of bare metal compute hosts using OpenStack ironic and ironic inspector
- Containerised workloads on bare metal using OpenStack magnum
- Big data on bare metal using OpenStack sahara