diff --git a/doc/source/deploy_install_guides/r4_release/bare_metal/aio_duplex.rst b/doc/source/deploy_install_guides/r4_release/bare_metal/aio_duplex.rst index 4b78e1d87..498e105aa 100644 --- a/doc/source/deploy_install_guides/r4_release/bare_metal/aio_duplex.rst +++ b/doc/source/deploy_install_guides/r4_release/bare_metal/aio_duplex.rst @@ -9,7 +9,7 @@ Overview .. include:: ../desc_aio_duplex.txt The bare metal AIO-DX deployment configuration may be extended with up to four -worker/compute nodes (not shown in the diagram). Installation instructions for +worker nodes (not shown in the diagram). Installation instructions for these additional nodes are described in :doc:`aio_duplex_extend`. .. include:: ../ipv6_note.txt diff --git a/doc/source/deploy_install_guides/r4_release/bare_metal/aio_duplex_extend.rst b/doc/source/deploy_install_guides/r4_release/bare_metal/aio_duplex_extend.rst index 946918d6d..2bba37680 100644 --- a/doc/source/deploy_install_guides/r4_release/bare_metal/aio_duplex_extend.rst +++ b/doc/source/deploy_install_guides/r4_release/bare_metal/aio_duplex_extend.rst @@ -1,26 +1,25 @@ -================================================ -Extend Capacity with Worker and/or Compute Nodes -================================================ +================================= +Extend Capacity with Worker Nodes +================================= -This section describes the steps to extend capacity with worker and/or compute -nodes on a **StarlingX R4.0 bare metal All-in-one Duplex** deployment -configuration. +This section describes the steps to extend capacity with worker nodes on a +**StarlingX R4.0 bare metal All-in-one Duplex** deployment configuration. .. contents:: :local: :depth: 1 ---------------------------------- -Install software on compute nodes ---------------------------------- +-------------------------------- +Install software on worker nodes +-------------------------------- -#. Power on the compute servers and force them to network boot with the +#. Power on the worker node servers and force them to network boot with the appropriate BIOS boot options for your particular server. -#. As the compute servers boot, a message appears on their console instructing +#. As the worker nodes boot, a message appears on their console instructing you to configure the personality of the node. -#. On the console of controller-0, list hosts to see newly discovered compute +#. On the console of controller-0, list hosts to see newly discovered worker node hosts (hostname=None): :: @@ -35,18 +34,19 @@ Install software on compute nodes | 4 | None | None | locked | disabled | offline | +----+--------------+-------------+----------------+-------------+--------------+ -#. Using the host id, set the personality of this host to 'controller': +#. Using the host id, set the personality of this host to 'worker': :: - system host-update 3 personality=worker hostname=compute-0 - system host-update 4 personality=worker hostname=compute-1 + system host-update 3 personality=worker hostname=worker-0 + system host-update 4 personality=worker hostname=worker-1 - This initiates the install of software on compute nodes. + This initiates the install of software on worker nodes. This can take 5-10 minutes, depending on the performance of the host machine. -#. Wait for the install of software on the computes to complete, the computes to - reboot and to both show as locked/disabled/online in 'system host-list'. +#. Wait for the install of software on the worker nodes to complete, for the + worker nodes to reboot, and for both to show as locked/disabled/online in + 'system host-list'. :: @@ -56,26 +56,26 @@ Install software on compute nodes +----+--------------+-------------+----------------+-------------+--------------+ | 1 | controller-0 | controller | unlocked | enabled | available | | 2 | controller-1 | controller | unlocked | enabled | available | - | 3 | compute-0 | compute | locked | disabled | online | - | 4 | compute-1 | compute | locked | disabled | online | + | 3 | worker-0 | worker | locked | disabled | online | + | 4 | worker-1 | worker | locked | disabled | online | +----+--------------+-------------+----------------+-------------+--------------+ ------------------------ -Configure compute nodes ------------------------ +---------------------- +Configure worker nodes +---------------------- -#. Assign the cluster-host network to the MGMT interface for the compute nodes: +#. Assign the cluster-host network to the MGMT interface for the worker nodes: (Note that the MGMT interfaces are partially set up automatically by the network install procedure.) :: - for COMPUTE in compute-0 compute-1; do - system interface-network-assign $COMPUTE mgmt0 cluster-host + for NODE in worker-0 worker-1; do + system interface-network-assign $NODE mgmt0 cluster-host done -#. Configure data interfaces for compute nodes. Use the DATA port names, for +#. Configure data interfaces for worker nodes. Use the DATA port names, for example eth0, that are applicable to your deployment environment. .. important:: @@ -117,11 +117,11 @@ Configure compute nodes system datanetwork-add ${PHYSNET0} vlan system datanetwork-add ${PHYSNET1} vlan - for COMPUTE in compute-0 compute-1; do - echo "Configuring interface for: $COMPUTE" + for NODE in worker-0 worker-1; do + echo "Configuring interface for: $NODE" set -ex - system host-port-list ${COMPUTE} --nowrap > ${SPL} - system host-if-list -a ${COMPUTE} --nowrap > ${SPIL} + system host-port-list ${NODE} --nowrap > ${SPL} + system host-if-list -a ${NODE} --nowrap > ${SPIL} DATA0PCIADDR=$(cat $SPL | grep $DATA0IF |awk '{print $8}') DATA1PCIADDR=$(cat $SPL | grep $DATA1IF |awk '{print $8}') DATA0PORTUUID=$(cat $SPL | grep ${DATA0PCIADDR} | awk '{print $2}') @@ -130,10 +130,10 @@ Configure compute nodes DATA1PORTNAME=$(cat $SPL | grep ${DATA1PCIADDR} | awk '{print $4}') DATA0IFUUID=$(cat $SPIL | awk -v DATA0PORTNAME=$DATA0PORTNAME '($12 ~ DATA0PORTNAME) {print $2}') DATA1IFUUID=$(cat $SPIL | awk -v DATA1PORTNAME=$DATA1PORTNAME '($12 ~ DATA1PORTNAME) {print $2}') - system host-if-modify -m 1500 -n data0 -c data ${COMPUTE} ${DATA0IFUUID} - system host-if-modify -m 1500 -n data1 -c data ${COMPUTE} ${DATA1IFUUID} - system interface-datanetwork-assign ${COMPUTE} ${DATA0IFUUID} ${PHYSNET0} - system interface-datanetwork-assign ${COMPUTE} ${DATA1IFUUID} ${PHYSNET1} + system host-if-modify -m 1500 -n data0 -c data ${NODE} ${DATA0IFUUID} + system host-if-modify -m 1500 -n data1 -c data ${NODE} ${DATA1IFUUID} + system interface-datanetwork-assign ${NODE} ${DATA0IFUUID} ${PHYSNET0} + system interface-datanetwork-assign ${NODE} ${DATA1IFUUID} ${PHYSNET1} set +ex done @@ -146,12 +146,12 @@ OpenStack-specific host configuration **This step is required only if the StarlingX OpenStack application (stx-openstack) will be installed.** -#. **For OpenStack only:** Assign OpenStack host labels to the compute nodes in +#. **For OpenStack only:** Assign OpenStack host labels to the worker nodes in support of installing the stx-openstack manifest and helm-charts later. :: - for NODE in compute-0 compute-1; do + for NODE in worker-0 worker-1; do system host-label-assign $NODE openstack-compute-node=enabled system host-label-assign $NODE openvswitch=enabled system host-label-assign $NODE sriov=enabled @@ -162,31 +162,31 @@ OpenStack-specific host configuration :: - for COMPUTE in compute-0 compute-1; do - echo "Configuring Nova local for: $COMPUTE" - ROOT_DISK=$(system host-show ${COMPUTE} | grep rootfs | awk '{print $4}') - ROOT_DISK_UUID=$(system host-disk-list ${COMPUTE} --nowrap | grep ${ROOT_DISK} | awk '{print $2}') + for NODE in worker-0 worker-1; do + echo "Configuring Nova local for: $NODE" + ROOT_DISK=$(system host-show ${NODE} | grep rootfs | awk '{print $4}') + ROOT_DISK_UUID=$(system host-disk-list ${NODE} --nowrap | grep ${ROOT_DISK} | awk '{print $2}') PARTITION_SIZE=10 - NOVA_PARTITION=$(system host-disk-partition-add -t lvm_phys_vol ${COMPUTE} ${ROOT_DISK_UUID} ${PARTITION_SIZE}) + NOVA_PARTITION=$(system host-disk-partition-add -t lvm_phys_vol ${NODE} ${ROOT_DISK_UUID} ${PARTITION_SIZE}) NOVA_PARTITION_UUID=$(echo ${NOVA_PARTITION} | grep -ow "| uuid | [a-z0-9\-]* |" | awk '{print $4}') - system host-lvg-add ${COMPUTE} nova-local - system host-pv-add ${COMPUTE} nova-local ${NOVA_PARTITION_UUID} + system host-lvg-add ${NODE} nova-local + system host-pv-add ${NODE} nova-local ${NOVA_PARTITION_UUID} done --------------------- -Unlock compute nodes --------------------- +------------------- +Unlock worker nodes +------------------- -Unlock compute nodes in order to bring them into service: +Unlock worker nodes in order to bring them into service: :: - for COMPUTE in compute-0 compute-1; do - system host-unlock $COMPUTE + for NODE in worker-0 worker-1; do + system host-unlock $NODE done -The compute nodes will reboot to apply configuration changes and come into +The worker nodes will reboot to apply configuration changes and come into service. This can take 5-10 minutes, depending on the performance of the host machine. diff --git a/doc/source/deploy_install_guides/r4_release/bare_metal/aio_duplex_install_kubernetes.rst b/doc/source/deploy_install_guides/r4_release/bare_metal/aio_duplex_install_kubernetes.rst index 5c733c6c1..0e1b1b918 100644 --- a/doc/source/deploy_install_guides/r4_release/bare_metal/aio_duplex_install_kubernetes.rst +++ b/doc/source/deploy_install_guides/r4_release/bare_metal/aio_duplex_install_kubernetes.rst @@ -212,13 +212,13 @@ Configure controller-0 DATA0IF= DATA1IF= - export COMPUTE=controller-0 + export NODE=controller-0 PHYSNET0='physnet0' PHYSNET1='physnet1' SPL=/tmp/tmp-system-port-list SPIL=/tmp/tmp-system-host-if-list - system host-port-list ${COMPUTE} --nowrap > ${SPL} - system host-if-list -a ${COMPUTE} --nowrap > ${SPIL} + system host-port-list ${NODE} --nowrap > ${SPL} + system host-if-list -a ${NODE} --nowrap > ${SPIL} DATA0PCIADDR=$(cat $SPL | grep $DATA0IF |awk '{print $8}') DATA1PCIADDR=$(cat $SPL | grep $DATA1IF |awk '{print $8}') DATA0PORTUUID=$(cat $SPL | grep ${DATA0PCIADDR} | awk '{print $2}') @@ -231,10 +231,10 @@ Configure controller-0 system datanetwork-add ${PHYSNET0} vlan system datanetwork-add ${PHYSNET1} vlan - system host-if-modify -m 1500 -n data0 -c data ${COMPUTE} ${DATA0IFUUID} - system host-if-modify -m 1500 -n data1 -c data ${COMPUTE} ${DATA1IFUUID} - system interface-datanetwork-assign ${COMPUTE} ${DATA0IFUUID} ${PHYSNET0} - system interface-datanetwork-assign ${COMPUTE} ${DATA1IFUUID} ${PHYSNET1} + system host-if-modify -m 1500 -n data0 -c data ${NODE} ${DATA0IFUUID} + system host-if-modify -m 1500 -n data1 -c data ${NODE} ${DATA1IFUUID} + system interface-datanetwork-assign ${NODE} ${DATA0IFUUID} ${PHYSNET0} + system interface-datanetwork-assign ${NODE} ${DATA1IFUUID} ${PHYSNET1} #. Add an OSD on controller-0 for Ceph. The following example adds an OSD to the `sdb` disk: @@ -362,13 +362,13 @@ Configure controller-1 DATA0IF= DATA1IF= - export COMPUTE=controller-1 + export NODE=controller-1 PHYSNET0='physnet0' PHYSNET1='physnet1' SPL=/tmp/tmp-system-port-list SPIL=/tmp/tmp-system-host-if-list - system host-port-list ${COMPUTE} --nowrap > ${SPL} - system host-if-list -a ${COMPUTE} --nowrap > ${SPIL} + system host-port-list ${NODE} --nowrap > ${SPL} + system host-if-list -a ${NODE} --nowrap > ${SPIL} DATA0PCIADDR=$(cat $SPL | grep $DATA0IF |awk '{print $8}') DATA1PCIADDR=$(cat $SPL | grep $DATA1IF |awk '{print $8}') DATA0PORTUUID=$(cat $SPL | grep ${DATA0PCIADDR} | awk '{print $2}') @@ -381,10 +381,10 @@ Configure controller-1 system datanetwork-add ${PHYSNET0} vlan system datanetwork-add ${PHYSNET1} vlan - system host-if-modify -m 1500 -n data0 -c data ${COMPUTE} ${DATA0IFUUID} - system host-if-modify -m 1500 -n data1 -c data ${COMPUTE} ${DATA1IFUUID} - system interface-datanetwork-assign ${COMPUTE} ${DATA0IFUUID} ${PHYSNET0} - system interface-datanetwork-assign ${COMPUTE} ${DATA1IFUUID} ${PHYSNET1} + system host-if-modify -m 1500 -n data0 -c data ${NODE} ${DATA0IFUUID} + system host-if-modify -m 1500 -n data1 -c data ${NODE} ${DATA1IFUUID} + system interface-datanetwork-assign ${NODE} ${DATA0IFUUID} ${PHYSNET0} + system interface-datanetwork-assign ${NODE} ${DATA1IFUUID} ${PHYSNET1} #. Add an OSD on controller-1 for Ceph: @@ -423,19 +423,19 @@ OpenStack-specific host configuration :: - export COMPUTE=controller-1 + export NODE=controller-1 echo ">>> Getting root disk info" - ROOT_DISK=$(system host-show ${COMPUTE} | grep rootfs | awk '{print $4}') - ROOT_DISK_UUID=$(system host-disk-list ${COMPUTE} --nowrap | grep ${ROOT_DISK} | awk '{print $2}') + ROOT_DISK=$(system host-show ${NODE} | grep rootfs | awk '{print $4}') + ROOT_DISK_UUID=$(system host-disk-list ${NODE} --nowrap | grep ${ROOT_DISK} | awk '{print $2}') echo "Root disk: $ROOT_DISK, UUID: $ROOT_DISK_UUID" echo ">>>> Configuring nova-local" NOVA_SIZE=34 - NOVA_PARTITION=$(system host-disk-partition-add -t lvm_phys_vol ${COMPUTE} ${ROOT_DISK_UUID} ${NOVA_SIZE}) + NOVA_PARTITION=$(system host-disk-partition-add -t lvm_phys_vol ${NODE} ${ROOT_DISK_UUID} ${NOVA_SIZE}) NOVA_PARTITION_UUID=$(echo ${NOVA_PARTITION} | grep -ow "| uuid | [a-z0-9\-]* |" | awk '{print $4}') - system host-lvg-add ${COMPUTE} nova-local - system host-pv-add ${COMPUTE} nova-local ${NOVA_PARTITION_UUID} + system host-lvg-add ${NODE} nova-local + system host-pv-add ${NODE} nova-local ${NOVA_PARTITION_UUID} sleep 2 ------------------- diff --git a/doc/source/deploy_install_guides/r4_release/bare_metal/aio_simplex_install_kubernetes.rst b/doc/source/deploy_install_guides/r4_release/bare_metal/aio_simplex_install_kubernetes.rst index edb654a77..1560acc58 100644 --- a/doc/source/deploy_install_guides/r4_release/bare_metal/aio_simplex_install_kubernetes.rst +++ b/doc/source/deploy_install_guides/r4_release/bare_metal/aio_simplex_install_kubernetes.rst @@ -158,8 +158,8 @@ Configure controller-0 source /etc/platform/openrc #. Configure the OAM interface of controller-0 and specify the attached network - as "oam". Use the OAM port name, for example eth0, that is applicable to your - deployment environment: + as "oam". Use the OAM port name that is applicable to your deployment + environment, for example eth0: :: @@ -219,13 +219,13 @@ Configure controller-0 DATA0IF= DATA1IF= - export COMPUTE=controller-0 + export NODE=controller-0 PHYSNET0='physnet0' PHYSNET1='physnet1' SPL=/tmp/tmp-system-port-list SPIL=/tmp/tmp-system-host-if-list - system host-port-list ${COMPUTE} --nowrap > ${SPL} - system host-if-list -a ${COMPUTE} --nowrap > ${SPIL} + system host-port-list ${NODE} --nowrap > ${SPL} + system host-if-list -a ${NODE} --nowrap > ${SPIL} DATA0PCIADDR=$(cat $SPL | grep $DATA0IF |awk '{print $8}') DATA1PCIADDR=$(cat $SPL | grep $DATA1IF |awk '{print $8}') DATA0PORTUUID=$(cat $SPL | grep ${DATA0PCIADDR} | awk '{print $2}') @@ -238,10 +238,10 @@ Configure controller-0 system datanetwork-add ${PHYSNET0} vlan system datanetwork-add ${PHYSNET1} vlan - system host-if-modify -m 1500 -n data0 -c data ${COMPUTE} ${DATA0IFUUID} - system host-if-modify -m 1500 -n data1 -c data ${COMPUTE} ${DATA1IFUUID} - system interface-datanetwork-assign ${COMPUTE} ${DATA0IFUUID} ${PHYSNET0} - system interface-datanetwork-assign ${COMPUTE} ${DATA1IFUUID} ${PHYSNET1} + system host-if-modify -m 1500 -n data0 -c data ${NODE} ${DATA0IFUUID} + system host-if-modify -m 1500 -n data1 -c data ${NODE} ${DATA1IFUUID} + system interface-datanetwork-assign ${NODE} ${DATA0IFUUID} ${PHYSNET0} + system interface-datanetwork-assign ${NODE} ${DATA1IFUUID} ${PHYSNET1} #. Add an OSD on controller-0 for Ceph. The following example adds an OSD to the `sdb` disk: @@ -286,7 +286,9 @@ OpenStack-specific host configuration manifest. * Shares the core(s) assigned to the platform. - If you require better performance, OVS-DPDK should be used: + If you require better performance, OVS-DPDK (OVS with the Data Plane + Development Kit, which is supported only on bare metal hardware) should be + used: * Runs directly on the host (it is not containerized). * Requires that at least 1 core be assigned/dedicated to the vSwitch function. @@ -300,8 +302,7 @@ OpenStack-specific host configuration Do not run any vSwitch directly on the host, instead, use the containerized OVS defined in the helm charts of stx-openstack manifest. - To deploy OVS-DPDK (OVS with the Data Plane Development Kit, which is - supported only on bare metal hardware), run the following command: + To deploy OVS-DPDK, run the following command: :: @@ -310,35 +311,60 @@ OpenStack-specific host configuration Once vswitch_type is set to OVS-DPDK, any subsequent nodes created will default to automatically assigning 1 vSwitch core for AIO controllers and 2 - vSwitch cores for computes. + vSwitch cores for compute-labeled worker nodes. - When using OVS-DPDK, virtual machines must be configured to use a flavor with - property: hw:mem_page_size=large + When using OVS-DPDK, configure vSwitch memory per NUMA node with the following + command: + + :: + + system host-memory-modify -f -1G <1G hugepages number> + + For example: + + :: + + system host-memory-modify -f vswitch -1G 1 worker-0 0 + + VMs created in an OVS-DPDK environment must be configured to use huge pages + to enable networking and must use a flavor with property: hw:mem_page_size=large + + Configure the huge pages for VMs in an OVS-DPDK environment with the command: + + :: + + system host-memory-modify -1G <1G hugepages number> + + For example: + + :: + + system host-memory-modify worker-0 0 -1G 10 .. note:: After controller-0 is unlocked, changing vswitch_type requires - locking and unlocking all computes (and/or AIO Controllers) to - apply the change. + locking and unlocking all compute-labeled worker nodes (and/or AIO + controllers) to apply the change. #. **For OpenStack only:** Set up disk partition for nova-local volume group, which is needed for stx-openstack nova ephemeral disks. :: - export COMPUTE=controller-0 + export NODE=controller-0 echo ">>> Getting root disk info" - ROOT_DISK=$(system host-show ${COMPUTE} | grep rootfs | awk '{print $4}') - ROOT_DISK_UUID=$(system host-disk-list ${COMPUTE} --nowrap | grep ${ROOT_DISK} | awk '{print $2}') + ROOT_DISK=$(system host-show ${NODE} | grep rootfs | awk '{print $4}') + ROOT_DISK_UUID=$(system host-disk-list ${NODE} --nowrap | grep ${ROOT_DISK} | awk '{print $2}') echo "Root disk: $ROOT_DISK, UUID: $ROOT_DISK_UUID" echo ">>>> Configuring nova-local" NOVA_SIZE=34 - NOVA_PARTITION=$(system host-disk-partition-add -t lvm_phys_vol ${COMPUTE} ${ROOT_DISK_UUID} ${NOVA_SIZE}) + NOVA_PARTITION=$(system host-disk-partition-add -t lvm_phys_vol ${NODE} ${ROOT_DISK_UUID} ${NOVA_SIZE}) NOVA_PARTITION_UUID=$(echo ${NOVA_PARTITION} | grep -ow "| uuid | [a-z0-9\-]* |" | awk '{print $4}') - system host-lvg-add ${COMPUTE} nova-local - system host-pv-add ${COMPUTE} nova-local ${NOVA_PARTITION_UUID} + system host-lvg-add ${NODE} nova-local + system host-pv-add ${NODE} nova-local ${NOVA_PARTITION_UUID} sleep 2 .. incl-config-controller-0-openstack-specific-aio-simplex-end: diff --git a/doc/source/deploy_install_guides/r4_release/bare_metal/controller_storage_hardware.rst b/doc/source/deploy_install_guides/r4_release/bare_metal/controller_storage_hardware.rst index f72ffab08..e1bb971a2 100644 --- a/doc/source/deploy_install_guides/r4_release/bare_metal/controller_storage_hardware.rst +++ b/doc/source/deploy_install_guides/r4_release/bare_metal/controller_storage_hardware.rst @@ -18,7 +18,7 @@ The recommended minimum hardware requirements for bare metal servers for various host types are: +-------------------------+-----------------------------+-----------------------------+ -| Minimum Requirement | Controller Node | Compute Node | +| Minimum Requirement | Controller Node | Worker Node | +=========================+=============================+=============================+ | Number of servers | 2 | 2-10 | +-------------------------+-----------------------------+-----------------------------+ diff --git a/doc/source/deploy_install_guides/r4_release/bare_metal/controller_storage_install_kubernetes.rst b/doc/source/deploy_install_guides/r4_release/bare_metal/controller_storage_install_kubernetes.rst index ba1016772..4b410ece4 100644 --- a/doc/source/deploy_install_guides/r4_release/bare_metal/controller_storage_install_kubernetes.rst +++ b/doc/source/deploy_install_guides/r4_release/bare_metal/controller_storage_install_kubernetes.rst @@ -231,7 +231,9 @@ OpenStack-specific host configuration manifest. * Shares the core(s) assigned to the platform. - If you require better performance, OVS-DPDK should be used: + If you require better performance, OVS-DPDK (OVS with the Data Plane + Development Kit, which is supported only on bare metal hardware) should be + used: * Runs directly on the host (it is not containerized). * Requires that at least 1 core be assigned/dedicated to the vSwitch function. @@ -245,26 +247,50 @@ OpenStack-specific host configuration Do not run any vSwitch directly on the host, instead, use the containerized OVS defined in the helm charts of stx-openstack manifest. - To deploy OVS-DPDK (OVS with the Data Plane Development Kit, which is - supported only on bare metal hardware), run the following command: + To deploy OVS-DPDK, run the following command: :: system modify --vswitch_type ovs-dpdk - system host-cpu-modify -f vswitch -p0 1 controller-0 + system host-cpu-modify -f vswitch -p0 1 controller-0 Once vswitch_type is set to OVS-DPDK, any subsequent nodes created will default to automatically assigning 1 vSwitch core for AIO controllers and 2 - vSwitch cores for computes. + vSwitch cores for compute-labeled worker nodes. - When using OVS-DPDK, Virtual Machines must be configured to use a flavor with - property: hw:mem_page_size=large. + When using OVS-DPDK, configure vSwitch memory per NUMA node with the following + command: + + :: + + system host-memory-modify -f -1G <1G hugepages number> + + For example: + + :: + + system host-memory-modify -f vswitch -1G 1 worker-0 0 + + VMs created in an OVS-DPDK environment must be configured to use huge pages + to enable networking and must use a flavor with property: hw:mem_page_size=large + + Configure the huge pages for VMs in an OVS-DPDK environment with the command: + + :: + + system host-memory-modify -1G <1G hugepages number> + + For example: + + :: + + system host-memory-modify worker-0 0 -1G 10 .. note:: - After controller-0 is unlocked, changing vswitch_type requires - locking and unlocking all computes (and/or AIO controllers) to - apply the change. + After controller-0 is unlocked, changing vswitch_type requires + locking and unlocking all compute-labeled worker nodes (and/or AIO + controllers) to apply the change. .. incl-config-controller-0-storage-end: @@ -281,9 +307,9 @@ Unlock controller-0 in order to bring it into service: Controller-0 will reboot in order to apply configuration changes and come into service. This can take 5-10 minutes, depending on the performance of the host machine. --------------------------------------------------- -Install software on controller-1 and compute nodes --------------------------------------------------- +------------------------------------------------- +Install software on controller-1 and worker nodes +------------------------------------------------- #. Power on the controller-1 server and force it to network boot with the appropriate BIOS boot options for your particular server. @@ -313,25 +339,24 @@ Install software on controller-1 and compute nodes This initiates the install of software on controller-1. This can take 5-10 minutes, depending on the performance of the host machine. -#. While waiting for the previous step to complete, power on the compute-0 and - compute-1 servers. Set the personality to 'worker' and assign a unique - hostname for each. +#. While waiting for the previous step to complete, power on the worker nodes. + Set the personality to 'worker' and assign a unique hostname for each. - For example, power on compute-0 and wait for the new host (hostname=None) to + For example, power on worker-0 and wait for the new host (hostname=None) to be discovered by checking 'system host-list': :: - system host-update 3 personality=worker hostname=compute-0 + system host-update 3 personality=worker hostname=worker-0 - Repeat for compute-1. Power on compute-1 and wait for the new host (hostname=None) to + Repeat for worker-1. Power on worker-1 and wait for the new host (hostname=None) to be discovered by checking 'system host-list': :: - system host-update 4 personality=worker hostname=compute-1 + system host-update 4 personality=worker hostname=worker-1-1 -#. Wait for the software installation on controller-1, compute-0, and compute-1 to +#. Wait for the software installation on controller-1, worker-0, and worker-1 to complete, for all servers to reboot, and for all to show as locked/disabled/online in 'system host-list'. @@ -344,8 +369,8 @@ Install software on controller-1 and compute nodes +----+--------------+-------------+----------------+-------------+--------------+ | 1 | controller-0 | controller | unlocked | enabled | available | | 2 | controller-1 | controller | locked | disabled | online | - | 3 | compute-0 | compute | locked | disabled | online | - | 4 | compute-1 | compute | locked | disabled | online | + | 3 | worker-0 | worker | locked | disabled | online | + | 4 | worker-1 | worker | locked | disabled | online | +----+--------------+-------------+----------------+-------------+--------------+ ---------------------- @@ -405,20 +430,20 @@ machine. .. incl-unlock-controller-1-end: ------------------------ -Configure compute nodes ------------------------ +---------------------- +Configure worker nodes +---------------------- -#. Add the third Ceph monitor to compute-0: +#. Add the third Ceph monitor to a worker node: (The first two Ceph monitors are automatically assigned to controller-0 and controller-1.) :: - system ceph-mon-add compute-0 + system ceph-mon-add worker-0 -#. Wait for the compute node monitor to complete configuration: +#. Wait for the worker node monitor to complete configuration: :: @@ -430,21 +455,21 @@ Configure compute nodes +--------------------------------------+-------+--------------+------------+------+ | 64176b6c-e284-4485-bb2a-115dee215279 | 20 | controller-1 | configured | None | | a9ca151b-7f2c-4551-8167-035d49e2df8c | 20 | controller-0 | configured | None | - | f76bc385-190c-4d9a-aa0f-107346a9907b | 20 | compute-0 | configured | None | + | f76bc385-190c-4d9a-aa0f-107346a9907b | 20 | worker-0 | configured | None | +--------------------------------------+-------+--------------+------------+------+ -#. Assign the cluster-host network to the MGMT interface for the compute nodes: +#. Assign the cluster-host network to the MGMT interface for the worker nodes: (Note that the MGMT interfaces are partially set up automatically by the network install procedure.) :: - for COMPUTE in compute-0 compute-1; do - system interface-network-assign $COMPUTE mgmt0 cluster-host + for NODE in worker-0 worker-1; do + system interface-network-assign $NODE mgmt0 cluster-host done -#. Configure data interfaces for compute nodes. Use the DATA port names, for +#. Configure data interfaces for worker nodes. Use the DATA port names, for example eth0, that are applicable to your deployment environment. .. important:: @@ -460,8 +485,8 @@ Configure compute nodes :: - for COMPUTE in compute-0 compute-1; do - system host-label-assign ${COMPUTE} sriovdp=enabled + for NODE in worker-0 worker-1; do + system host-label-assign ${NODE} sriovdp=enabled done * If planning on running DPDK in containers on this host, configure the number @@ -469,9 +494,9 @@ Configure compute nodes :: - for COMPUTE in compute-0 compute-1; do - system host-memory-modify ${COMPUTE} 0 -1G 100 - system host-memory-modify ${COMPUTE} 1 -1G 100 + for NODE in worker-0 worker-1; do + system host-memory-modify ${NODE} 0 -1G 100 + system host-memory-modify ${NODE} 1 -1G 100 done For both Kubernetes and OpenStack: @@ -490,11 +515,11 @@ Configure compute nodes system datanetwork-add ${PHYSNET0} vlan system datanetwork-add ${PHYSNET1} vlan - for COMPUTE in compute-0 compute-1; do - echo "Configuring interface for: $COMPUTE" + for NODE in worker-0 worker-1; do + echo "Configuring interface for: $NODE" set -ex - system host-port-list ${COMPUTE} --nowrap > ${SPL} - system host-if-list -a ${COMPUTE} --nowrap > ${SPIL} + system host-port-list ${NODE} --nowrap > ${SPL} + system host-if-list -a ${NODE} --nowrap > ${SPIL} DATA0PCIADDR=$(cat $SPL | grep $DATA0IF |awk '{print $8}') DATA1PCIADDR=$(cat $SPL | grep $DATA1IF |awk '{print $8}') DATA0PORTUUID=$(cat $SPL | grep ${DATA0PCIADDR} | awk '{print $2}') @@ -503,10 +528,10 @@ Configure compute nodes DATA1PORTNAME=$(cat $SPL | grep ${DATA1PCIADDR} | awk '{print $4}') DATA0IFUUID=$(cat $SPIL | awk -v DATA0PORTNAME=$DATA0PORTNAME '($12 ~ DATA0PORTNAME) {print $2}') DATA1IFUUID=$(cat $SPIL | awk -v DATA1PORTNAME=$DATA1PORTNAME '($12 ~ DATA1PORTNAME) {print $2}') - system host-if-modify -m 1500 -n data0 -c data ${COMPUTE} ${DATA0IFUUID} - system host-if-modify -m 1500 -n data1 -c data ${COMPUTE} ${DATA1IFUUID} - system interface-datanetwork-assign ${COMPUTE} ${DATA0IFUUID} ${PHYSNET0} - system interface-datanetwork-assign ${COMPUTE} ${DATA1IFUUID} ${PHYSNET1} + system host-if-modify -m 1500 -n data0 -c data ${NODE} ${DATA0IFUUID} + system host-if-modify -m 1500 -n data1 -c data ${NODE} ${DATA1IFUUID} + system interface-datanetwork-assign ${NODE} ${DATA0IFUUID} ${PHYSNET0} + system interface-datanetwork-assign ${NODE} ${DATA1IFUUID} ${PHYSNET1} set +ex done @@ -519,12 +544,12 @@ OpenStack-specific host configuration **This step is required only if the StarlingX OpenStack application (stx-openstack) will be installed.** -#. **For OpenStack only:** Assign OpenStack host labels to the compute nodes in +#. **For OpenStack only:** Assign OpenStack host labels to the worker nodes in support of installing the stx-openstack manifest and helm-charts later. :: - for NODE in compute-0 compute-1; do + for NODE in worker-0 worker-1; do system host-label-assign $NODE openstack-compute-node=enabled system host-label-assign $NODE openvswitch=enabled system host-label-assign $NODE sriov=enabled @@ -535,30 +560,30 @@ OpenStack-specific host configuration :: - for COMPUTE in compute-0 compute-1; do - echo "Configuring Nova local for: $COMPUTE" - ROOT_DISK=$(system host-show ${COMPUTE} | grep rootfs | awk '{print $4}') - ROOT_DISK_UUID=$(system host-disk-list ${COMPUTE} --nowrap | grep ${ROOT_DISK} | awk '{print $2}') + for NODE in worker-0 worker-1; do + echo "Configuring Nova local for: $NODE" + ROOT_DISK=$(system host-show ${NODE} | grep rootfs | awk '{print $4}') + ROOT_DISK_UUID=$(system host-disk-list ${NODE} --nowrap | grep ${ROOT_DISK} | awk '{print $2}') PARTITION_SIZE=10 - NOVA_PARTITION=$(system host-disk-partition-add -t lvm_phys_vol ${COMPUTE} ${ROOT_DISK_UUID} ${PARTITION_SIZE}) + NOVA_PARTITION=$(system host-disk-partition-add -t lvm_phys_vol ${NODE} ${ROOT_DISK_UUID} ${PARTITION_SIZE}) NOVA_PARTITION_UUID=$(echo ${NOVA_PARTITION} | grep -ow "| uuid | [a-z0-9\-]* |" | awk '{print $4}') - system host-lvg-add ${COMPUTE} nova-local - system host-pv-add ${COMPUTE} nova-local ${NOVA_PARTITION_UUID} + system host-lvg-add ${NODE} nova-local + system host-pv-add ${NODE} nova-local ${NOVA_PARTITION_UUID} done -------------------- -Unlock compute nodes +Unlock worker nodes -------------------- -Unlock compute nodes in order to bring them into service: +Unlock worker nodes in order to bring them into service: :: - for COMPUTE in compute-0 compute-1; do - system host-unlock $COMPUTE + for NODE in worker-0 worker-1; do + system host-unlock $NODE done -The compute nodes will reboot in order to apply configuration changes and come into +The worker nodes will reboot in order to apply configuration changes and come into service. This can take 5-10 minutes, depending on the performance of the host machine. ---------------------------- diff --git a/doc/source/deploy_install_guides/r4_release/bare_metal/dedicated_storage_hardware.rst b/doc/source/deploy_install_guides/r4_release/bare_metal/dedicated_storage_hardware.rst index f1f80032d..46a1127ec 100644 --- a/doc/source/deploy_install_guides/r4_release/bare_metal/dedicated_storage_hardware.rst +++ b/doc/source/deploy_install_guides/r4_release/bare_metal/dedicated_storage_hardware.rst @@ -18,7 +18,7 @@ The recommended minimum hardware requirements for bare metal servers for various host types are: +---------------------+---------------------------+-----------------------+-----------------------+ -| Minimum Requirement | Controller Node | Storage Node | Compute Node | +| Minimum Requirement | Controller Node | Storage Node | Worker Node | +=====================+===========================+=======================+=======================+ | Number of servers | 2 | 2-9 | 2-100 | +---------------------+---------------------------+-----------------------+-----------------------+ diff --git a/doc/source/deploy_install_guides/r4_release/bare_metal/dedicated_storage_install_kubernetes.rst b/doc/source/deploy_install_guides/r4_release/bare_metal/dedicated_storage_install_kubernetes.rst index 98365da36..ea91844dd 100644 --- a/doc/source/deploy_install_guides/r4_release/bare_metal/dedicated_storage_install_kubernetes.rst +++ b/doc/source/deploy_install_guides/r4_release/bare_metal/dedicated_storage_install_kubernetes.rst @@ -60,9 +60,9 @@ Unlock controller-0 in order to bring it into service: Controller-0 will reboot in order to apply configuration changes and come into service. This can take 5-10 minutes, depending on the performance of the host machine. ------------------------------------------------------------------- -Install software on controller-1, storage nodes, and compute nodes ------------------------------------------------------------------- +----------------------------------------------------------------- +Install software on controller-1, storage nodes, and worker nodes +----------------------------------------------------------------- #. Power on the controller-1 server and force it to network boot with the appropriate BIOS boot options for your particular server. @@ -113,28 +113,27 @@ Install software on controller-1, storage nodes, and compute nodes This initiates the software installation on storage-0 and storage-1. This can take 5-10 minutes, depending on the performance of the host machine. -#. While waiting for the previous step to complete, power on the compute-0 and - compute-1 servers. Set the personality to 'worker' and assign a unique - hostname for each. +#. While waiting for the previous step to complete, power on the worker nodes. + Set the personality to 'worker' and assign a unique hostname for each. - For example, power on compute-0 and wait for the new host (hostname=None) to + For example, power on worker-0 and wait for the new host (hostname=None) to be discovered by checking 'system host-list': :: - system host-update 5 personality=worker hostname=compute-0 + system host-update 5 personality=worker hostname=worker-0 - Repeat for compute-1. Power on compute-1 and wait for the new host + Repeat for worker-1. Power on worker-1 and wait for the new host (hostname=None) to be discovered by checking 'system host-list': :: - system host-update 6 personality=worker hostname=compute-1 + system host-update 6 personality=worker hostname=worker-1 - This initiates the install of software on compute-0 and compute-1. + This initiates the install of software on worker-0 and worker-1. #. Wait for the software installation on controller-1, storage-0, storage-1, - compute-0, and compute-1 to complete, for all servers to reboot, and for all to + worker-0, and worker-1 to complete, for all servers to reboot, and for all to show as locked/disabled/online in 'system host-list'. :: @@ -147,8 +146,8 @@ Install software on controller-1, storage nodes, and compute nodes | 2 | controller-1 | controller | locked | disabled | online | | 3 | storage-0 | storage | locked | disabled | online | | 4 | storage-1 | storage | locked | disabled | online | - | 5 | compute-0 | compute | locked | disabled | online | - | 6 | compute-1 | compute | locked | disabled | online | + | 5 | worker-0 | worker | locked | disabled | online | + | 6 | worker-1 | worker | locked | disabled | online | +----+--------------+-------------+----------------+-------------+--------------+ ---------------------- @@ -178,8 +177,8 @@ Configure storage nodes :: - for COMPUTE in storage-0 storage-1; do - system interface-network-assign $COMPUTE mgmt0 cluster-host + for NODE in storage-0 storage-1; do + system interface-network-assign $NODE mgmt0 cluster-host done #. Add OSDs to storage-0. The following example adds OSDs to the `sdb` disk: @@ -228,22 +227,22 @@ The storage nodes will reboot in order to apply configuration changes and come into service. This can take 5-10 minutes, depending on the performance of the host machine. ------------------------ -Configure compute nodes ------------------------ +---------------------- +Configure worker nodes +---------------------- -#. Assign the cluster-host network to the MGMT interface for the compute nodes: +#. Assign the cluster-host network to the MGMT interface for the worker nodes: (Note that the MGMT interfaces are partially set up automatically by the network install procedure.) :: - for COMPUTE in compute-0 compute-1; do - system interface-network-assign $COMPUTE mgmt0 cluster-host + for NODE in worker-0 worker-1; do + system interface-network-assign $NODE mgmt0 cluster-host done -#. Configure data interfaces for compute nodes. Use the DATA port names, for +#. Configure data interfaces for worker nodes. Use the DATA port names, for example eth0, that are applicable to your deployment environment. .. important:: @@ -259,8 +258,8 @@ Configure compute nodes :: - for COMPUTE in compute-0 compute-1; do - system host-label-assign ${COMPUTE} sriovdp=enabled + for NODE in worker-0 worker-1; do + system host-label-assign ${NODE} sriovdp=enabled done * If planning on running DPDK in containers on this host, configure the number @@ -268,9 +267,9 @@ Configure compute nodes :: - for COMPUTE in compute-0 compute-1; do - system host-memory-modify ${COMPUTE} 0 -1G 100 - system host-memory-modify ${COMPUTE} 1 -1G 100 + for NODE in worker-0 worker-1; do + system host-memory-modify ${NODE} 0 -1G 100 + system host-memory-modify ${NODE} 1 -1G 100 done For both Kubernetes and OpenStack: @@ -289,11 +288,11 @@ Configure compute nodes system datanetwork-add ${PHYSNET0} vlan system datanetwork-add ${PHYSNET1} vlan - for COMPUTE in compute-0 compute-1; do - echo "Configuring interface for: $COMPUTE" + for NODE in worker-0 worker-1; do + echo "Configuring interface for: $NODE" set -ex - system host-port-list ${COMPUTE} --nowrap > ${SPL} - system host-if-list -a ${COMPUTE} --nowrap > ${SPIL} + system host-port-list ${NODE} --nowrap > ${SPL} + system host-if-list -a ${NODE} --nowrap > ${SPIL} DATA0PCIADDR=$(cat $SPL | grep $DATA0IF |awk '{print $8}') DATA1PCIADDR=$(cat $SPL | grep $DATA1IF |awk '{print $8}') DATA0PORTUUID=$(cat $SPL | grep ${DATA0PCIADDR} | awk '{print $2}') @@ -302,10 +301,10 @@ Configure compute nodes DATA1PORTNAME=$(cat $SPL | grep ${DATA1PCIADDR} | awk '{print $4}') DATA0IFUUID=$(cat $SPIL | awk -v DATA0PORTNAME=$DATA0PORTNAME '($12 ~ DATA0PORTNAME) {print $2}') DATA1IFUUID=$(cat $SPIL | awk -v DATA1PORTNAME=$DATA1PORTNAME '($12 ~ DATA1PORTNAME) {print $2}') - system host-if-modify -m 1500 -n data0 -c data ${COMPUTE} ${DATA0IFUUID} - system host-if-modify -m 1500 -n data1 -c data ${COMPUTE} ${DATA1IFUUID} - system interface-datanetwork-assign ${COMPUTE} ${DATA0IFUUID} ${PHYSNET0} - system interface-datanetwork-assign ${COMPUTE} ${DATA1IFUUID} ${PHYSNET1} + system host-if-modify -m 1500 -n data0 -c data ${NODE} ${DATA0IFUUID} + system host-if-modify -m 1500 -n data1 -c data ${NODE} ${DATA1IFUUID} + system interface-datanetwork-assign ${NODE} ${DATA0IFUUID} ${PHYSNET0} + system interface-datanetwork-assign ${NODE} ${DATA1IFUUID} ${PHYSNET1} set +ex done @@ -318,12 +317,12 @@ OpenStack-specific host configuration **This step is required only if the StarlingX OpenStack application (stx-openstack) will be installed.** -#. **For OpenStack only:** Assign OpenStack host labels to the compute nodes in +#. **For OpenStack only:** Assign OpenStack host labels to the worker nodes in support of installing the stx-openstack manifest and helm-charts later. :: - for NODE in compute-0 compute-1; do + for NODE in worker-0 worker-1; do system host-label-assign $NODE openstack-compute-node=enabled system host-label-assign $NODE openvswitch=enabled system host-label-assign $NODE sriov=enabled @@ -334,30 +333,30 @@ OpenStack-specific host configuration :: - for COMPUTE in compute-0 compute-1; do - echo "Configuring Nova local for: $COMPUTE" - ROOT_DISK=$(system host-show ${COMPUTE} | grep rootfs | awk '{print $4}') - ROOT_DISK_UUID=$(system host-disk-list ${COMPUTE} --nowrap | grep ${ROOT_DISK} | awk '{print $2}') + for NODE in worker-0 worker-1; do + echo "Configuring Nova local for: $NODE" + ROOT_DISK=$(system host-show ${NODE} | grep rootfs | awk '{print $4}') + ROOT_DISK_UUID=$(system host-disk-list ${NODE} --nowrap | grep ${ROOT_DISK} | awk '{print $2}') PARTITION_SIZE=10 - NOVA_PARTITION=$(system host-disk-partition-add -t lvm_phys_vol ${COMPUTE} ${ROOT_DISK_UUID} ${PARTITION_SIZE}) + NOVA_PARTITION=$(system host-disk-partition-add -t lvm_phys_vol ${NODE} ${ROOT_DISK_UUID} ${PARTITION_SIZE}) NOVA_PARTITION_UUID=$(echo ${NOVA_PARTITION} | grep -ow "| uuid | [a-z0-9\-]* |" | awk '{print $4}') - system host-lvg-add ${COMPUTE} nova-local - system host-pv-add ${COMPUTE} nova-local ${NOVA_PARTITION_UUID} + system host-lvg-add ${NODE} nova-local + system host-pv-add ${NODE} nova-local ${NOVA_PARTITION_UUID} done --------------------- -Unlock compute nodes --------------------- +------------------- +Unlock worker nodes +------------------- -Unlock compute nodes in order to bring them into service: +Unlock worker nodes in order to bring them into service: :: - for COMPUTE in compute-0 compute-1; do - system host-unlock $COMPUTE + for NODE in worker-0 worker-1; do + system host-unlock $NODE done -The compute nodes will reboot in order to apply configuration changes and come +The worker nodes will reboot in order to apply configuration changes and come into service. This can take 5-10 minutes, depending on the performance of the host machine. diff --git a/doc/source/deploy_install_guides/r4_release/bare_metal/ironic.rst b/doc/source/deploy_install_guides/r4_release/bare_metal/ironic.rst index 92c93285c..063119c51 100644 --- a/doc/source/deploy_install_guides/r4_release/bare_metal/ironic.rst +++ b/doc/source/deploy_install_guides/r4_release/bare_metal/ironic.rst @@ -56,8 +56,8 @@ standard configuration, either: This guide assumes that you have a standard deployment installed and configured -with 2x controllers and at least 1x compute node, with the StarlingX OpenStack -application (stx-openstack) applied. +with 2x controllers and at least 1x compute-labeled worker node, with the +StarlingX OpenStack application (stx-openstack) applied. .. toctree:: :maxdepth: 1 diff --git a/doc/source/deploy_install_guides/r4_release/bare_metal/ironic_hardware.rst b/doc/source/deploy_install_guides/r4_release/bare_metal/ironic_hardware.rst index 4bebfc023..ffee52ab3 100644 --- a/doc/source/deploy_install_guides/r4_release/bare_metal/ironic_hardware.rst +++ b/doc/source/deploy_install_guides/r4_release/bare_metal/ironic_hardware.rst @@ -23,15 +23,15 @@ For controller nodes: * Additional NIC port on both controller nodes for connecting to the ironic-provisioning-net. -For compute nodes: +For worker nodes: * If using a flat data network for the Ironic provisioning network, an additional - NIC port on one of the compute nodes is required. + NIC port on one of the worker nodes is required. * Alternatively, use a VLAN data network for the Ironic provisioning network and - simply add the new data network to an existing interface on the compute node. + simply add the new data network to an existing interface on the worker node. -* Additional switch ports / configuration for new ports on controller, compute, +* Additional switch ports / configuration for new ports on controller, worker, and Ironic nodes, for connectivity to the Ironic provisioning network. ----------------------------------- diff --git a/doc/source/deploy_install_guides/r4_release/bare_metal/ironic_install.rst b/doc/source/deploy_install_guides/r4_release/bare_metal/ironic_install.rst index fb5ae7b5a..e07ce7a09 100644 --- a/doc/source/deploy_install_guides/r4_release/bare_metal/ironic_install.rst +++ b/doc/source/deploy_install_guides/r4_release/bare_metal/ironic_install.rst @@ -113,14 +113,14 @@ Configure Ironic network on deployed standard StarlingX # Unlock the node to apply changes system host-unlock controller-1 -#. Configure the new interface (for Ironic) on one of the compute nodes and - assign it to the Ironic data network. This example uses the data network - `ironic-data` that was named in a previous step. +#. Configure the new interface (for Ironic) on one of the compute-labeled worker + nodes and assign it to the Ironic data network. This example uses the data + network `ironic-data` that was named in a previous step. :: - system interface-datanetwork-assign compute-0 eno1 ironic-data - system host-if-modify -n ironicdata -c data compute-0 eno1 + system interface-datanetwork-assign worker-0 eno1 ironic-data + system host-if-modify -n ironicdata -c data worker-0 eno1 **************************** Generate user Helm overrides diff --git a/doc/source/deploy_install_guides/r4_release/desc_aio_duplex.txt b/doc/source/deploy_install_guides/r4_release/desc_aio_duplex.txt index 952f5836c..079d2af47 100644 --- a/doc/source/deploy_install_guides/r4_release/desc_aio_duplex.txt +++ b/doc/source/deploy_install_guides/r4_release/desc_aio_duplex.txt @@ -1,16 +1,16 @@ The All-in-one Duplex (AIO-DX) deployment option provides a pair of high availability (HA) servers with each server providing all three cloud functions -(controller, compute, and storage). +(controller, worker, and storage). An AIO-DX configuration provides the following benefits: * Only a small amount of cloud processing and storage power is required -* Application consolidation using multiple virtual machines on a single pair of - physical servers +* Application consolidation using multiple containers or virtual machines on a + single pair of physical servers * High availability (HA) services run on the controller function across two physical servers in either active/active or active/standby mode * A storage back end solution using a two-node CEPH deployment across two servers -* Virtual machines scheduled on both compute functions +* Containers or virtual machines scheduled on both worker functions * Protection against overall server hardware fault, where * All controller HA services go active on the remaining healthy server diff --git a/doc/source/deploy_install_guides/r4_release/desc_aio_simplex.txt b/doc/source/deploy_install_guides/r4_release/desc_aio_simplex.txt index b3bd342ef..5f10318b2 100644 --- a/doc/source/deploy_install_guides/r4_release/desc_aio_simplex.txt +++ b/doc/source/deploy_install_guides/r4_release/desc_aio_simplex.txt @@ -1,10 +1,10 @@ The All-in-one Simplex (AIO-SX) deployment option provides all three cloud -functions (controller, compute, and storage) on a single server with the +functions (controller, worker, and storage) on a single server with the following benefits: * Requires only a small amount of cloud processing and storage power -* Application consolidation using multiple virtual machines on a single pair of - physical servers +* Application consolidation using multiple containers or virtual machines on a + single pair of physical servers * A storage backend solution using a single-node CEPH deployment .. figure:: ../figures/starlingx-deployment-options-simplex.png diff --git a/doc/source/deploy_install_guides/r4_release/desc_controller_storage.txt b/doc/source/deploy_install_guides/r4_release/desc_controller_storage.txt index cf53c35ad..8a952326e 100644 --- a/doc/source/deploy_install_guides/r4_release/desc_controller_storage.txt +++ b/doc/source/deploy_install_guides/r4_release/desc_controller_storage.txt @@ -1,19 +1,19 @@ The Standard with Controller Storage deployment option provides two high -availability (HA) controller nodes and a pool of up to 10 compute nodes. +availability (HA) controller nodes and a pool of up to 10 worker nodes. A Standard with Controller Storage configuration provides the following benefits: -* A pool of up to 10 compute nodes +* A pool of up to 10 worker nodes * High availability (HA) services run across the controller nodes in either active/active or active/standby mode * A storage back end solution using a two-node CEPH deployment across two controller servers -* Protection against overall controller and compute node failure, where +* Protection against overall controller and worker node failure, where * On overall controller node failure, all controller HA services go active on the remaining healthy controller node - * On overall compute node failure, virtual machines and containers are - recovered on the remaining healthy compute nodes + * On overall worker node failure, virtual machines and containers are + recovered on the remaining healthy worker nodes .. figure:: ../figures/starlingx-deployment-options-controller-storage.png :scale: 50% diff --git a/doc/source/deploy_install_guides/r4_release/desc_dedicated_storage.txt b/doc/source/deploy_install_guides/r4_release/desc_dedicated_storage.txt index ef41856fd..c352663af 100644 --- a/doc/source/deploy_install_guides/r4_release/desc_dedicated_storage.txt +++ b/doc/source/deploy_install_guides/r4_release/desc_dedicated_storage.txt @@ -1,9 +1,9 @@ The Standard with Dedicated Storage deployment option is a standard installation -with independent controller, compute, and storage nodes. +with independent controller, worker, and storage nodes. A Standard with Dedicated Storage configuration provides the following benefits: -* A pool of up to 100 compute nodes +* A pool of up to 100 worker nodes * A 2x node high availability (HA) controller cluster with HA services running across the controller nodes in either active/active or active/standby mode * A storage back end solution using a two-to-9x node HA CEPH storage cluster diff --git a/doc/source/deploy_install_guides/r4_release/distributed_cloud/index.rst b/doc/source/deploy_install_guides/r4_release/distributed_cloud/index.rst index 6a1b3c55e..edb1d2f76 100644 --- a/doc/source/deploy_install_guides/r4_release/distributed_cloud/index.rst +++ b/doc/source/deploy_install_guides/r4_release/distributed_cloud/index.rst @@ -204,9 +204,9 @@ At the subcloud location: RTNETLINK answers: File exists Adding route to System Controller... DONE -At the SystemController: +At the System Controller: -1. Create a ``bootstrap-values.yml`` overrides file for the subcloud, for +1. Create a ``bootstrap-values.yml`` override file for the subcloud. For example: .. code:: yaml @@ -215,15 +215,21 @@ At the SystemController: name: "subcloud1" description: "Ottawa Site" location: "YOW" + management_subnet: 192.168.101.0/24 management_start_address: 192.168.101.2 management_end_address: 192.168.101.50 management_gateway_address: 192.168.101.1 + external_oam_subnet: 10.10.10.0/24 external_oam_gateway_address: 10.10.10.1 external_oam_floating_address: 10.10.10.12 + systemcontroller_gateway_address: 192.168.204.101 + .. important:: The `management_*` entries in the override file are required + for all installation types, including AIO-Simplex. + 2. Add the subcloud using the CLI command below: .. code:: sh @@ -273,7 +279,10 @@ At the SystemController: - For Standard with dedicated storage nodes: `Bare metal Standard with Dedicated Storage Installation R4.0 `_ -5. Add routes from the subcloud to the controller management network: +On the active controller for each subcloud: + +#. Add a route from the subcloud to the controller management network to enable + the subcloud to go online. For each host in the subcloud: .. code:: sh @@ -286,4 +295,3 @@ At the SystemController: system host-route-add 1 enp0s8 192.168.204.0 24 192.168.101.1 - Repeat this step for each host of the subcloud. diff --git a/doc/source/deploy_install_guides/r4_release/openstack/install.rst b/doc/source/deploy_install_guides/r4_release/openstack/install.rst index 7f66478f4..979f54862 100644 --- a/doc/source/deploy_install_guides/r4_release/openstack/install.rst +++ b/doc/source/deploy_install_guides/r4_release/openstack/install.rst @@ -15,21 +15,17 @@ StarlingX Kubernetes platform: Install application manifest and helm-charts -------------------------------------------- -#. Get the StarlingX OpenStack application (stx-openstack) manifest and helm-charts. - This can be from a private StarlingX build or, as shown below, from the public - Cengen StarlingX build off ``master`` branch: - - :: - - wget http://mirror.starlingx.cengn.ca/mirror/starlingx/release/2.0.0/centos/outputs/helm-charts/stx-openstack-1.0-17-centos-stable-latest.tgz +#. Get the latest StarlingX OpenStack application (stx-openstack) manifest and + helm-charts. This can be from a private StarlingX build or from the public + `CENGN StarlingX mirror `_. #. Load the stx-openstack application's package into StarlingX. The tarball package contains stx-openstack's Airship Armada manifest and stx-openstack's - set of helm charts: + set of helm charts. For example: :: - system application-upload stx-openstack-1.0-17-centos-stable-latest.tgz + system application-upload stx-openstack--centos-stable-latest.tgz This will: diff --git a/doc/source/deploy_install_guides/r4_release/virtual/aio_duplex_environ.rst b/doc/source/deploy_install_guides/r4_release/virtual/aio_duplex_environ.rst index 6f31dc257..e099bee12 100644 --- a/doc/source/deploy_install_guides/r4_release/virtual/aio_duplex_environ.rst +++ b/doc/source/deploy_install_guides/r4_release/virtual/aio_duplex_environ.rst @@ -19,10 +19,6 @@ Physical host requirements and setup Prepare virtual environment and servers --------------------------------------- -The following steps explain how to prepare the virtual environment and servers -on a physical host for a StarlingX R4.0 virtual All-in-one Duplex deployment -configuration. - #. Prepare virtual environment. Set up the virtual platform networks for virtual deployment: diff --git a/doc/source/deploy_install_guides/r4_release/virtual/aio_duplex_install_kubernetes.rst b/doc/source/deploy_install_guides/r4_release/virtual/aio_duplex_install_kubernetes.rst index d9c5d5d40..b6cf566ed 100644 --- a/doc/source/deploy_install_guides/r4_release/virtual/aio_duplex_install_kubernetes.rst +++ b/doc/source/deploy_install_guides/r4_release/virtual/aio_duplex_install_kubernetes.rst @@ -206,13 +206,13 @@ On virtual controller-0: DATA0IF=eth1000 DATA1IF=eth1001 - export COMPUTE=controller-0 + export NODE=controller-0 PHYSNET0='physnet0' PHYSNET1='physnet1' SPL=/tmp/tmp-system-port-list SPIL=/tmp/tmp-system-host-if-list - system host-port-list ${COMPUTE} --nowrap > ${SPL} - system host-if-list -a ${COMPUTE} --nowrap > ${SPIL} + system host-port-list ${NODE} --nowrap > ${SPL} + system host-if-list -a ${NODE} --nowrap > ${SPIL} DATA0PCIADDR=$(cat $SPL | grep $DATA0IF |awk '{print $8}') DATA1PCIADDR=$(cat $SPL | grep $DATA1IF |awk '{print $8}') DATA0PORTUUID=$(cat $SPL | grep ${DATA0PCIADDR} | awk '{print $2}') @@ -225,10 +225,10 @@ On virtual controller-0: system datanetwork-add ${PHYSNET0} vlan system datanetwork-add ${PHYSNET1} vlan - system host-if-modify -m 1500 -n data0 -c data ${COMPUTE} ${DATA0IFUUID} - system host-if-modify -m 1500 -n data1 -c data ${COMPUTE} ${DATA1IFUUID} - system interface-datanetwork-assign ${COMPUTE} ${DATA0IFUUID} ${PHYSNET0} - system interface-datanetwork-assign ${COMPUTE} ${DATA1IFUUID} ${PHYSNET1} + system host-if-modify -m 1500 -n data0 -c data ${NODE} ${DATA0IFUUID} + system host-if-modify -m 1500 -n data1 -c data ${NODE} ${DATA1IFUUID} + system interface-datanetwork-assign ${NODE} ${DATA0IFUUID} ${PHYSNET0} + system interface-datanetwork-assign ${NODE} ${DATA1IFUUID} ${PHYSNET1} #. Add an OSD on controller-0 for Ceph: @@ -351,13 +351,13 @@ On virtual controller-0: DATA0IF=eth1000 DATA1IF=eth1001 - export COMPUTE=controller-1 + export NODE=controller-1 PHYSNET0='physnet0' PHYSNET1='physnet1' SPL=/tmp/tmp-system-port-list SPIL=/tmp/tmp-system-host-if-list - system host-port-list ${COMPUTE} --nowrap > ${SPL} - system host-if-list -a ${COMPUTE} --nowrap > ${SPIL} + system host-port-list ${NODE} --nowrap > ${SPL} + system host-if-list -a ${NODE} --nowrap > ${SPIL} DATA0PCIADDR=$(cat $SPL | grep $DATA0IF |awk '{print $8}') DATA1PCIADDR=$(cat $SPL | grep $DATA1IF |awk '{print $8}') DATA0PORTUUID=$(cat $SPL | grep ${DATA0PCIADDR} | awk '{print $2}') @@ -370,10 +370,10 @@ On virtual controller-0: system datanetwork-add ${PHYSNET0} vlan system datanetwork-add ${PHYSNET1} vlan - system host-if-modify -m 1500 -n data0 -c data ${COMPUTE} ${DATA0IFUUID} - system host-if-modify -m 1500 -n data1 -c data ${COMPUTE} ${DATA1IFUUID} - system interface-datanetwork-assign ${COMPUTE} ${DATA0IFUUID} ${PHYSNET0} - system interface-datanetwork-assign ${COMPUTE} ${DATA1IFUUID} ${PHYSNET1} + system host-if-modify -m 1500 -n data0 -c data ${NODE} ${DATA0IFUUID} + system host-if-modify -m 1500 -n data1 -c data ${NODE} ${DATA1IFUUID} + system interface-datanetwork-assign ${NODE} ${DATA0IFUUID} ${PHYSNET0} + system interface-datanetwork-assign ${NODE} ${DATA1IFUUID} ${PHYSNET1} #. Add an OSD on controller-1 for Ceph: @@ -412,19 +412,19 @@ OpenStack-specific host configuration :: - export COMPUTE=controller-1 + export NODE=controller-1 echo ">>> Getting root disk info" - ROOT_DISK=$(system host-show ${COMPUTE} | grep rootfs | awk '{print $4}') - ROOT_DISK_UUID=$(system host-disk-list ${COMPUTE} --nowrap | grep ${ROOT_DISK} | awk '{print $2}') + ROOT_DISK=$(system host-show ${NODE} | grep rootfs | awk '{print $4}') + ROOT_DISK_UUID=$(system host-disk-list ${NODE} --nowrap | grep ${ROOT_DISK} | awk '{print $2}') echo "Root disk: $ROOT_DISK, UUID: $ROOT_DISK_UUID" echo ">>>> Configuring nova-local" NOVA_SIZE=34 - NOVA_PARTITION=$(system host-disk-partition-add -t lvm_phys_vol ${COMPUTE} ${ROOT_DISK_UUID} ${NOVA_SIZE}) + NOVA_PARTITION=$(system host-disk-partition-add -t lvm_phys_vol ${NODE} ${ROOT_DISK_UUID} ${NOVA_SIZE}) NOVA_PARTITION_UUID=$(echo ${NOVA_PARTITION} | grep -ow "| uuid | [a-z0-9\-]* |" | awk '{print $4}') - system host-lvg-add ${COMPUTE} nova-local - system host-pv-add ${COMPUTE} nova-local ${NOVA_PARTITION_UUID} + system host-lvg-add ${NODE} nova-local + system host-pv-add ${NODE} nova-local ${NOVA_PARTITION_UUID} ------------------- Unlock controller-1 diff --git a/doc/source/deploy_install_guides/r4_release/virtual/aio_simplex_environ.rst b/doc/source/deploy_install_guides/r4_release/virtual/aio_simplex_environ.rst index d03ece814..50ca70e9d 100644 --- a/doc/source/deploy_install_guides/r4_release/virtual/aio_simplex_environ.rst +++ b/doc/source/deploy_install_guides/r4_release/virtual/aio_simplex_environ.rst @@ -19,10 +19,6 @@ Physical host requirements and setup Prepare virtual environment and servers --------------------------------------- -The following steps explain how to prepare the virtual environment and servers -on a physical host for a StarlingX R4.0 virtual All-in-one Simplex deployment -configuration. - #. Prepare virtual environment. Set up the virtual platform networks for virtual deployment: diff --git a/doc/source/deploy_install_guides/r4_release/virtual/aio_simplex_install_kubernetes.rst b/doc/source/deploy_install_guides/r4_release/virtual/aio_simplex_install_kubernetes.rst index 8f7d1ce2e..e966686e3 100644 --- a/doc/source/deploy_install_guides/r4_release/virtual/aio_simplex_install_kubernetes.rst +++ b/doc/source/deploy_install_guides/r4_release/virtual/aio_simplex_install_kubernetes.rst @@ -197,13 +197,13 @@ On virtual controller-0: DATA0IF=eth1000 DATA1IF=eth1001 - export COMPUTE=controller-0 + export NODE=controller-0 PHYSNET0='physnet0' PHYSNET1='physnet1' SPL=/tmp/tmp-system-port-list SPIL=/tmp/tmp-system-host-if-list - system host-port-list ${COMPUTE} --nowrap > ${SPL} - system host-if-list -a ${COMPUTE} --nowrap > ${SPIL} + system host-port-list ${NODE} --nowrap > ${SPL} + system host-if-list -a ${NODE} --nowrap > ${SPIL} DATA0PCIADDR=$(cat $SPL | grep $DATA0IF |awk '{print $8}') DATA1PCIADDR=$(cat $SPL | grep $DATA1IF |awk '{print $8}') DATA0PORTUUID=$(cat $SPL | grep ${DATA0PCIADDR} | awk '{print $2}') @@ -216,10 +216,10 @@ On virtual controller-0: system datanetwork-add ${PHYSNET0} vlan system datanetwork-add ${PHYSNET1} vlan - system host-if-modify -m 1500 -n data0 -c data ${COMPUTE} ${DATA0IFUUID} - system host-if-modify -m 1500 -n data1 -c data ${COMPUTE} ${DATA1IFUUID} - system interface-datanetwork-assign ${COMPUTE} ${DATA0IFUUID} ${PHYSNET0} - system interface-datanetwork-assign ${COMPUTE} ${DATA1IFUUID} ${PHYSNET1} + system host-if-modify -m 1500 -n data0 -c data ${NODE} ${DATA0IFUUID} + system host-if-modify -m 1500 -n data1 -c data ${NODE} ${DATA1IFUUID} + system interface-datanetwork-assign ${NODE} ${DATA0IFUUID} ${PHYSNET0} + system interface-datanetwork-assign ${NODE} ${DATA1IFUUID} ${PHYSNET1} #. Add an OSD on controller-0 for Ceph: @@ -267,19 +267,19 @@ OpenStack-specific host configuration :: - export COMPUTE=controller-0 + export NODE=controller-0 echo ">>> Getting root disk info" - ROOT_DISK=$(system host-show ${COMPUTE} | grep rootfs | awk '{print $4}') - ROOT_DISK_UUID=$(system host-disk-list ${COMPUTE} --nowrap | grep ${ROOT_DISK} | awk '{print $2}') + ROOT_DISK=$(system host-show ${NODE} | grep rootfs | awk '{print $4}') + ROOT_DISK_UUID=$(system host-disk-list ${NODE} --nowrap | grep ${ROOT_DISK} | awk '{print $2}') echo "Root disk: $ROOT_DISK, UUID: $ROOT_DISK_UUID" echo ">>>> Configuring nova-local" NOVA_SIZE=34 - NOVA_PARTITION=$(system host-disk-partition-add -t lvm_phys_vol ${COMPUTE} ${ROOT_DISK_UUID} ${NOVA_SIZE}) + NOVA_PARTITION=$(system host-disk-partition-add -t lvm_phys_vol ${NODE} ${ROOT_DISK_UUID} ${NOVA_SIZE}) NOVA_PARTITION_UUID=$(echo ${NOVA_PARTITION} | grep -ow "| uuid | [a-z0-9\-]* |" | awk '{print $4}') - system host-lvg-add ${COMPUTE} nova-local - system host-pv-add ${COMPUTE} nova-local ${NOVA_PARTITION_UUID} + system host-lvg-add ${NODE} nova-local + system host-pv-add ${NODE} nova-local ${NOVA_PARTITION_UUID} sleep 2 .. incl-config-controller-0-openstack-specific-aio-simplex-end: diff --git a/doc/source/deploy_install_guides/r4_release/virtual/controller_storage_environ.rst b/doc/source/deploy_install_guides/r4_release/virtual/controller_storage_environ.rst index 18fe9941c..890dd4d6f 100644 --- a/doc/source/deploy_install_guides/r4_release/virtual/controller_storage_environ.rst +++ b/doc/source/deploy_install_guides/r4_release/virtual/controller_storage_environ.rst @@ -20,10 +20,6 @@ Physical host requirements and setup Prepare virtual environment and servers --------------------------------------- -The following steps explain how to prepare the virtual environment and servers -on a physical host for a StarlingX R4.0 virtual Standard with Controller Storage -deployment configuration. - #. Prepare virtual environment. Set up virtual platform networks for virtual deployment: diff --git a/doc/source/deploy_install_guides/r4_release/virtual/controller_storage_install_kubernetes.rst b/doc/source/deploy_install_guides/r4_release/virtual/controller_storage_install_kubernetes.rst index b434b4560..41332a8ee 100644 --- a/doc/source/deploy_install_guides/r4_release/virtual/controller_storage_install_kubernetes.rst +++ b/doc/source/deploy_install_guides/r4_release/virtual/controller_storage_install_kubernetes.rst @@ -237,9 +237,9 @@ Unlock virtual controller-0 in order to bring it into service: Controller-0 will reboot in order to apply configuration changes and come into service. This can take 5-10 minutes, depending on the performance of the host machine. --------------------------------------------------- -Install software on controller-1 and compute nodes --------------------------------------------------- +------------------------------------------------- +Install software on controller-1 and worker nodes +------------------------------------------------- #. On the host, power on the controller-1 virtual server, 'controllerstorage-controller-1'. It will automatically attempt to network @@ -296,7 +296,7 @@ Install software on controller-1 and compute nodes :: - system host-update 3 personality=worker hostname=compute-0 + system host-update 3 personality=worker hostname=worker-0 Repeat for 'controllerstorage-worker-1'. On the host: @@ -309,9 +309,9 @@ Install software on controller-1 and compute nodes :: - system host-update 4 personality=worker hostname=compute-1 + system host-update 4 personality=worker hostname=worker-1 -#. Wait for the software installation on controller-1, compute-0, and compute-1 to +#. Wait for the software installation on controller-1, worker-0, and worker-1 to complete, for all virtual servers to reboot, and for all to show as locked/disabled/online in 'system host-list'. @@ -323,8 +323,8 @@ Install software on controller-1 and compute nodes +----+--------------+-------------+----------------+-------------+--------------+ | 1 | controller-0 | controller | unlocked | enabled | available | | 2 | controller-1 | controller | locked | disabled | online | - | 3 | compute-0 | compute | locked | disabled | online | - | 4 | compute-1 | compute | locked | disabled | online | + | 3 | worker-0 | worker | locked | disabled | online | + | 4 | worker-1 | worker | locked | disabled | online | +----+--------------+-------------+----------------+-------------+--------------+ ---------------------- @@ -379,22 +379,22 @@ service. This can take 5-10 minutes, depending on the performance of the host ma .. incl-unlock-controller-1-virt-controller-storage-end: ------------------------ -Configure compute nodes ------------------------ +---------------------- +Configure worker nodes +---------------------- On virtual controller-0: -#. Add the third Ceph monitor to compute-0: +#. Add the third Ceph monitor to a worker node: (The first two Ceph monitors are automatically assigned to controller-0 and controller-1.) :: - system ceph-mon-add compute-0 + system ceph-mon-add worker-0 -#. Wait for the compute node monitor to complete configuration: +#. Wait for the worker node monitor to complete configuration: :: @@ -406,21 +406,21 @@ On virtual controller-0: +--------------------------------------+-------+--------------+------------+------+ | 64176b6c-e284-4485-bb2a-115dee215279 | 20 | controller-1 | configured | None | | a9ca151b-7f2c-4551-8167-035d49e2df8c | 20 | controller-0 | configured | None | - | f76bc385-190c-4d9a-aa0f-107346a9907b | 20 | compute-0 | configured | None | + | f76bc385-190c-4d9a-aa0f-107346a9907b | 20 | worker-0 | configured | None | +--------------------------------------+-------+--------------+------------+------+ -#. Assign the cluster-host network to the MGMT interface for the compute nodes. +#. Assign the cluster-host network to the MGMT interface for the worker nodes. Note that the MGMT interfaces are partially set up automatically by the network install procedure. :: - for COMPUTE in compute-0 compute-1; do - system interface-network-assign $COMPUTE mgmt0 cluster-host + for NODE in worker-0 worker-1; do + system interface-network-assign $NODE mgmt0 cluster-host done -#. Configure data interfaces for compute nodes. +#. Configure data interfaces for worker nodes. .. important:: @@ -447,11 +447,11 @@ On virtual controller-0: system datanetwork-add ${PHYSNET0} vlan system datanetwork-add ${PHYSNET1} vlan - for COMPUTE in compute-0 compute-1; do - echo "Configuring interface for: $COMPUTE" + for NODE in worker-0 worker-1; do + echo "Configuring interface for: $NODE" set -ex - system host-port-list ${COMPUTE} --nowrap > ${SPL} - system host-if-list -a ${COMPUTE} --nowrap > ${SPIL} + system host-port-list ${NODE} --nowrap > ${SPL} + system host-if-list -a ${NODE} --nowrap > ${SPIL} DATA0PCIADDR=$(cat $SPL | grep $DATA0IF |awk '{print $8}') DATA1PCIADDR=$(cat $SPL | grep $DATA1IF |awk '{print $8}') DATA0PORTUUID=$(cat $SPL | grep ${DATA0PCIADDR} | awk '{print $2}') @@ -460,10 +460,10 @@ On virtual controller-0: DATA1PORTNAME=$(cat $SPL | grep ${DATA1PCIADDR} | awk '{print $4}') DATA0IFUUID=$(cat $SPIL | awk -v DATA0PORTNAME=$DATA0PORTNAME '($12 ~ DATA0PORTNAME) {print $2}') DATA1IFUUID=$(cat $SPIL | awk -v DATA1PORTNAME=$DATA1PORTNAME '($12 ~ DATA1PORTNAME) {print $2}') - system host-if-modify -m 1500 -n data0 -c data ${COMPUTE} ${DATA0IFUUID} - system host-if-modify -m 1500 -n data1 -c data ${COMPUTE} ${DATA1IFUUID} - system interface-datanetwork-assign ${COMPUTE} ${DATA0IFUUID} ${PHYSNET0} - system interface-datanetwork-assign ${COMPUTE} ${DATA1IFUUID} ${PHYSNET1} + system host-if-modify -m 1500 -n data0 -c data ${NODE} ${DATA0IFUUID} + system host-if-modify -m 1500 -n data1 -c data ${NODE} ${DATA1IFUUID} + system interface-datanetwork-assign ${NODE} ${DATA0IFUUID} ${PHYSNET0} + system interface-datanetwork-assign ${NODE} ${DATA1IFUUID} ${PHYSNET1} set +ex done @@ -476,12 +476,12 @@ OpenStack-specific host configuration **This step is required only if the StarlingX OpenStack application (stx-openstack) will be installed.** -#. **For OpenStack only:** Assign OpenStack host labels to the compute nodes in +#. **For OpenStack only:** Assign OpenStack host labels to the worker nodes in support of installing the stx-openstack manifest/helm-charts later: :: - for NODE in compute-0 compute-1; do + for NODE in worker-0 worker-1; do system host-label-assign $NODE openstack-compute-node=enabled system host-label-assign $NODE openvswitch=enabled system host-label-assign $NODE sriov=enabled @@ -492,32 +492,32 @@ OpenStack-specific host configuration :: - for COMPUTE in compute-0 compute-1; do - echo "Configuring Nova local for: $COMPUTE" - ROOT_DISK=$(system host-show ${COMPUTE} | grep rootfs | awk '{print $4}') - ROOT_DISK_UUID=$(system host-disk-list ${COMPUTE} --nowrap | grep ${ROOT_DISK} | awk '{print $2}') + for NODE in worker-0 worker-1; do + echo "Configuring Nova local for: $NODE" + ROOT_DISK=$(system host-show ${NODE} | grep rootfs | awk '{print $4}') + ROOT_DISK_UUID=$(system host-disk-list ${NODE} --nowrap | grep ${ROOT_DISK} | awk '{print $2}') PARTITION_SIZE=10 - NOVA_PARTITION=$(system host-disk-partition-add -t lvm_phys_vol ${COMPUTE} ${ROOT_DISK_UUID} ${PARTITION_SIZE}) + NOVA_PARTITION=$(system host-disk-partition-add -t lvm_phys_vol ${NODE} ${ROOT_DISK_UUID} ${PARTITION_SIZE}) NOVA_PARTITION_UUID=$(echo ${NOVA_PARTITION} | grep -ow "| uuid | [a-z0-9\-]* |" | awk '{print $4}') - system host-lvg-add ${COMPUTE} nova-local - system host-pv-add ${COMPUTE} nova-local ${NOVA_PARTITION_UUID} + system host-lvg-add ${NODE} nova-local + system host-pv-add ${NODE} nova-local ${NOVA_PARTITION_UUID} done --------------------- -Unlock compute nodes --------------------- +------------------- +Unlock worker nodes +------------------- .. incl-unlock-compute-nodes-virt-controller-storage-start: -Unlock virtual compute nodes to bring them into service: +Unlock virtual worker nodes to bring them into service: :: - for COMPUTE in compute-0 compute-1; do - system host-unlock $COMPUTE + for NODE in worker-0 worker-1; do + system host-unlock $NODE done -The compute nodes will reboot in order to apply configuration changes and come into +The worker nodes will reboot in order to apply configuration changes and come into service. This can take 5-10 minutes, depending on the performance of the host machine. .. incl-unlock-compute-nodes-virt-controller-storage-end: diff --git a/doc/source/deploy_install_guides/r4_release/virtual/dedicated_storage_environ.rst b/doc/source/deploy_install_guides/r4_release/virtual/dedicated_storage_environ.rst index 256172c28..eb2fb5552 100644 --- a/doc/source/deploy_install_guides/r4_release/virtual/dedicated_storage_environ.rst +++ b/doc/source/deploy_install_guides/r4_release/virtual/dedicated_storage_environ.rst @@ -20,10 +20,6 @@ Physical host requirements and setup Prepare virtual environment and servers --------------------------------------- -The following steps explain how to prepare the virtual environment and servers -on a physical host for a StarlingX R4.0 virtual Standard with Dedicated Storage -deployment configuration. - #. Prepare virtual environment. Set up virtual platform networks for virtual deployment: diff --git a/doc/source/deploy_install_guides/r4_release/virtual/dedicated_storage_install_kubernetes.rst b/doc/source/deploy_install_guides/r4_release/virtual/dedicated_storage_install_kubernetes.rst index 953dc67b7..d4dd6bc76 100644 --- a/doc/source/deploy_install_guides/r4_release/virtual/dedicated_storage_install_kubernetes.rst +++ b/doc/source/deploy_install_guides/r4_release/virtual/dedicated_storage_install_kubernetes.rst @@ -77,9 +77,9 @@ Unlock virtual controller-0 in order to bring it into service: Controller-0 will reboot in order to apply configuration changes and come into service. This can take 5-10 minutes, depending on the performance of the host machine. ------------------------------------------------------------------- -Install software on controller-1, storage nodes, and compute nodes ------------------------------------------------------------------- +----------------------------------------------------------------- +Install software on controller-1, storage nodes, and worker nodes +----------------------------------------------------------------- #. On the host, power on the controller-1 virtual server, 'dedicatedstorage-controller-1'. It will automatically attempt to network @@ -168,7 +168,7 @@ Install software on controller-1, storage nodes, and compute nodes :: - system host-update 5 personality=worker hostname=compute-0 + system host-update 5 personality=worker hostname=worker-0 Repeat for 'dedicatedstorage-worker-1'. On the host: @@ -181,12 +181,12 @@ Install software on controller-1, storage nodes, and compute nodes :: - ssystem host-update 6 personality=worker hostname=compute-1 + ssystem host-update 6 personality=worker hostname=worker-1 - This initiates software installation on compute-0 and compute-1. + This initiates software installation on worker-0 and worker-1. #. Wait for the software installation on controller-1, storage-0, storage-1, - compute-0, and compute-1 to complete, for all virtual servers to reboot, and for all + worker-0, and worker-1 to complete, for all virtual servers to reboot, and for all to show as locked/disabled/online in 'system host-list'. :: @@ -199,8 +199,8 @@ Install software on controller-1, storage nodes, and compute nodes | 2 | controller-1 | controller | locked | disabled | online | | 3 | storage-0 | storage | locked | disabled | online | | 4 | storage-1 | storage | locked | disabled | online | - | 5 | compute-0 | compute | locked | disabled | online | - | 6 | compute-1 | compute | locked | disabled | online | + | 5 | worker-0 | worker | locked | disabled | online | + | 6 | worker-1 | worker | locked | disabled | online | +----+--------------+-------------+----------------+-------------+--------------+ ---------------------- @@ -231,8 +231,8 @@ On virtual controller-0: :: - for COMPUTE in storage-0 storage-1; do - system interface-network-assign $COMPUTE mgmt0 cluster-host + for NODE in storage-0 storage-1; do + system interface-network-assign $NODE mgmt0 cluster-host done #. Add OSDs to storage-0: @@ -278,24 +278,24 @@ Unlock virtual storage nodes in order to bring them into service: The storage nodes will reboot in order to apply configuration changes and come into service. This can take 5-10 minutes, depending on the performance of the host machine. ------------------------ -Configure compute nodes ------------------------ +---------------------- +Configure worker nodes +---------------------- On virtual controller-0: -#. Assign the cluster-host network to the MGMT interface for the compute nodes. +#. Assign the cluster-host network to the MGMT interface for the worker nodes. Note that the MGMT interfaces are partially set up automatically by the network install procedure. :: - for COMPUTE in compute-0 compute-1; do - system interface-network-assign $COMPUTE mgmt0 cluster-host + for NODE in worker-0 worker-1; do + system interface-network-assign $NODE mgmt0 cluster-host done -#. Configure data interfaces for compute nodes. +#. Configure data interfaces for worker nodes. .. important:: @@ -325,11 +325,11 @@ On virtual controller-0: system datanetwork-add ${PHYSNET0} vlan system datanetwork-add ${PHYSNET1} vlan - for COMPUTE in compute-0 compute-1; do - echo "Configuring interface for: $COMPUTE" + for NODE in worker-0 worker-1; do + echo "Configuring interface for: $NODE" set -ex - system host-port-list ${COMPUTE} --nowrap > ${SPL} - system host-if-list -a ${COMPUTE} --nowrap > ${SPIL} + system host-port-list ${NODE} --nowrap > ${SPL} + system host-if-list -a ${NODE} --nowrap > ${SPIL} DATA0PCIADDR=$(cat $SPL | grep $DATA0IF |awk '{print $8}') DATA1PCIADDR=$(cat $SPL | grep $DATA1IF |awk '{print $8}') DATA0PORTUUID=$(cat $SPL | grep ${DATA0PCIADDR} | awk '{print $2}') @@ -338,10 +338,10 @@ On virtual controller-0: DATA1PORTNAME=$(cat $SPL | grep ${DATA1PCIADDR} | awk '{print $4}') DATA0IFUUID=$(cat $SPIL | awk -v DATA0PORTNAME=$DATA0PORTNAME '($12 ~ DATA0PORTNAME) {print $2}') DATA1IFUUID=$(cat $SPIL | awk -v DATA1PORTNAME=$DATA1PORTNAME '($12 ~ DATA1PORTNAME) {print $2}') - system host-if-modify -m 1500 -n data0 -c data ${COMPUTE} ${DATA0IFUUID} - system host-if-modify -m 1500 -n data1 -c data ${COMPUTE} ${DATA1IFUUID} - system interface-datanetwork-assign ${COMPUTE} ${DATA0IFUUID} ${PHYSNET0} - system interface-datanetwork-assign ${COMPUTE} ${DATA1IFUUID} ${PHYSNET1} + system host-if-modify -m 1500 -n data0 -c data ${NODE} ${DATA0IFUUID} + system host-if-modify -m 1500 -n data1 -c data ${NODE} ${DATA1IFUUID} + system interface-datanetwork-assign ${NODE} ${DATA0IFUUID} ${PHYSNET0} + system interface-datanetwork-assign ${NODE} ${DATA1IFUUID} ${PHYSNET1} set +ex done @@ -354,12 +354,12 @@ OpenStack-specific host configuration **This step is required only if the StarlingX OpenStack application (stx-openstack) will be installed.** -#. **For OpenStack only:** Assign OpenStack host labels to the compute nodes in +#. **For OpenStack only:** Assign OpenStack host labels to the worker nodes in support of installing the stx-openstack manifest/helm-charts later: :: - for NODE in compute-0 compute-1; do + for NODE in worker-0 worker-1; do system host-label-assign $NODE openstack-compute-node=enabled system host-label-assign $NODE openvswitch=enabled system host-label-assign $NODE sriov=enabled @@ -370,20 +370,20 @@ OpenStack-specific host configuration :: - for COMPUTE in compute-0 compute-1; do - echo "Configuring Nova local for: $COMPUTE" - ROOT_DISK=$(system host-show ${COMPUTE} | grep rootfs | awk '{print $4}') - ROOT_DISK_UUID=$(system host-disk-list ${COMPUTE} --nowrap | grep ${ROOT_DISK} | awk '{print $2}') + for NODE in worker-0 worker-1; do + echo "Configuring Nova local for: $NODE" + ROOT_DISK=$(system host-show ${NODE} | grep rootfs | awk '{print $4}') + ROOT_DISK_UUID=$(system host-disk-list ${NODE} --nowrap | grep ${ROOT_DISK} | awk '{print $2}') PARTITION_SIZE=10 - NOVA_PARTITION=$(system host-disk-partition-add -t lvm_phys_vol ${COMPUTE} ${ROOT_DISK_UUID} ${PARTITION_SIZE}) + NOVA_PARTITION=$(system host-disk-partition-add -t lvm_phys_vol ${NODE} ${ROOT_DISK_UUID} ${PARTITION_SIZE}) NOVA_PARTITION_UUID=$(echo ${NOVA_PARTITION} | grep -ow "| uuid | [a-z0-9\-]* |" | awk '{print $4}') - system host-lvg-add ${COMPUTE} nova-local - system host-pv-add ${COMPUTE} nova-local ${NOVA_PARTITION_UUID} + system host-lvg-add ${NODE} nova-local + system host-pv-add ${NODE} nova-local ${NOVA_PARTITION_UUID} done --------------------- -Unlock compute nodes --------------------- +------------------- +Unlock worker nodes +------------------- .. include:: controller_storage_install_kubernetes.rst :start-after: incl-unlock-compute-nodes-virt-controller-storage-start: diff --git a/doc/source/deploy_install_guides/r4_release/virtual/physical_host_req.txt b/doc/source/deploy_install_guides/r4_release/virtual/physical_host_req.txt index db257f386..422a8696b 100644 --- a/doc/source/deploy_install_guides/r4_release/virtual/physical_host_req.txt +++ b/doc/source/deploy_install_guides/r4_release/virtual/physical_host_req.txt @@ -67,9 +67,6 @@ Set up the host with the following steps: apparmor-profile modules. -#. Get the StarlingX ISO. This can be from a private StarlingX build or from the public Cengn - StarlingX build off 'master' branch, as shown below: - - :: - - wget http://mirror.starlingx.cengn.ca/mirror/starlingx/release/2.0.0/centos/outputs/iso/bootimage.iso +#. Get the StarlingX ISO from the + `CENGN StarlingX mirror `_. + Alternately, you can use an ISO from a private StarlingX build.