openstack-ansible/deploy-guide/source/targethosts-prepare.rst
Dmitriy Rabotyagov 3c8b6d199c [doc] Update Ubuntu 22.04 support status
We should not consider Ubuntu Jammy deployment as experimental anymore.
The only issue we have with it right now is version selection of ceph
packages, that is only one of many possible storage drivers.

We also update overview-requirements to have only relevant information.

Change-Id: Ide828700bba5c2a3e8d49e4ea3591f2d458e7519
2023-02-08 10:15:48 +01:00

5.8 KiB

Configuring the operating system

This section describes the installation and configuration of operating systems for the target hosts, as well as deploying SSH keys and configuring storage.

Installing the operating system

Install one of the following supported operating systems on the target host:

  • Ubuntu server 22.04 (Focal Fossa) LTS 64-bit Note: Due to the absence of community repos, Ceph packages will be installed from the default repository which contains only Quincy release.
  • Ubuntu server 20.04 (Focal Fossa) LTS 64-bit
  • Debian 11 64-bit
  • Centos 9 Stream 64-bit
  • Rocky Linux 9 64-bit

Configure at least one network interface to access the Internet or suitable local repositories.

Some distributions add an extraneous entry in the /etc/hosts file that resolves the actual hostname to another loopback IP address such as 127.0.1.1. You must comment out or remove this entry to prevent name resolution problems. Do not remove the 127.0.0.1 entry. This step is especially important for metal deployments.

We recommend adding the Secure Shell (SSH) server packages to the installation on target hosts that do not have local (console) access.

Note

We also recommend setting your locale to en_US.UTF-8. Other locales might work, but they are not tested or supported.

Configure Debian

  1. Update package source lists

    # apt update
  2. Upgrade the system packages and kernel:

    # apt dist-upgrade
  3. Install additional software packages:

    # apt install bridge-utils debootstrap ifenslave ifenslave-2.6 \
      lsof lvm2 openssh-server sudo tcpdump vlan python3
  4. Reboot the host to activate the changes and use the new kernel.

Configure Ubuntu

  1. Update package source lists

    # apt update
  2. Upgrade the system packages and kernel:

    # apt dist-upgrade
  3. Install additional software packages:

    # apt install bridge-utils debootstrap openssh-server \
      tcpdump vlan python3
  4. Install the kernel extra package if you have one for your kernel version .. code-block:: shell-session

    # apt install linux-modules-extra-$(uname -r)

  5. Reboot the host to activate the changes and use the new kernel.

Configure CentOS / Rocky

  1. Upgrade the system packages and kernel:

    # dnf upgrade
  2. Disable SELinux. Edit /etc/sysconfig/selinux, make sure that SELINUX=enforcing is changed to SELINUX=disabled.

    Note

    SELinux enabled is not currently supported in OpenStack-Ansible for CentOS/RHEL due to a lack of maintainers for the feature.

  3. Install additional software packages:

    # dnf install iputils lsof openssh-server\
      sudo tcpdump python3
  4. (Optional) Reduce the kernel log level by changing the printk value in your sysctls:

    # echo "kernel.printk='4 1 7 4'" >> /etc/sysctl.conf
  5. Reboot the host to activate the changes and use the new kernel.

Configure SSH keys

Ansible uses SSH to connect the deployment host and target hosts.

  1. Copy the contents of the public key file on the deployment host to the /root/.ssh/authorized_keys file on each target host.
  2. Test public key authentication from the deployment host to each target host by using SSH to connect to the target host from the deployment host. If you can connect and get the shell without authenticating, it is working. SSH provides a shell without asking for a password.

For more information about how to generate an SSH key pair, as well as best practices, see GitHub's documentation about generating SSH keys.

Important

OpenStack-Ansible deployments require the presence of a /root/.ssh/id_rsa.pub file on the deployment host. The contents of this file is inserted into an authorized_keys file for the containers, which is a necessary step for the Ansible playbooks. You can override this behavior by setting the lxc_container_ssh_key variable to the public key for the container.

Configuring the storage

Logical Volume Manager (LVM) enables a single device to be split into multiple logical volumes that appear as a physical storage device to the operating system. The Block Storage (cinder) service, and LXC containers that optionally run the OpenStack infrastructure, can optionally use LVM for their data storage.

Note

OpenStack-Ansible automatically configures LVM on the nodes, and overrides any existing LVM configuration. If you had a customized LVM configuration, edit the generated configuration file as needed.

  1. To use the optional Block Storage (cinder) service, create an LVM volume group named cinder-volumes on the storage host. Specify a metadata size of 2048 when creating the physical volume. For example:

    # pvcreate --metadatasize 2048 physical_volume_device_path
    # vgcreate cinder-volumes physical_volume_device_path
  2. Optionally, create an LVM volume group named lxc for container file systems and set lxc_container_backing_store: lvm in user_variables.yml if you want to use LXC with LVM. If the lxc volume group does not exist, containers are automatically installed on the file system under /var/lib/lxc by default.