Merge "Remove spurious escapes (r8,dsR8)"

This commit is contained in:
Zuul 2023-03-15 19:51:09 +00:00 committed by Gerrit Code Review
commit dfa7f00983
358 changed files with 1273 additions and 1270 deletions

View File

@ -6,8 +6,8 @@
appropriate settings.
In particular, if |IGMP| snooping is enabled on |ToR| switches, then a
device acting as an |IGMP| querier is required on the network \(on the same
|VLAN|\) to prevent nodes from being dropped from the multicast group. The
device acting as an |IGMP| querier is required on the network (on the same
|VLAN|) to prevent nodes from being dropped from the multicast group. The
|IGMP| querier periodically sends |IGMP| queries to all nodes on the
network, and each node sends an |IGMP| join or report in response. Without
an |IGMP| querier, the nodes do not periodically send |IGMP| join messages

View File

@ -1,8 +1,8 @@
.. begin-redfish-vms
For subclouds with servers that support Redfish Virtual Media Service \(version
1.2 or higher\), you can use the Central Cloud's CLI to install the ISO and
For subclouds with servers that support Redfish Virtual Media Service (version
1.2 or higher), you can use the Central Cloud's CLI to install the ISO and
bootstrap subclouds from the Central Cloud. For more information, see
:ref:`Installing a Subcloud Using Redfish Platform Management Service
<installing-a-subcloud-using-redfish-platform-management-service>`.

View File

@ -20,7 +20,7 @@ more letters, as follows:
A slash-separated list of letters is used when the alarm can be triggered with
one of several severity levels.
An asterisk \(\*\) indicates the management-affecting severity, if any. A
An asterisk (\*) indicates the management-affecting severity, if any. A
management-affecting alarm is one that cannot be ignored at the indicated
severity level or higher by using relaxed alarm rules during an orchestrated
patch or upgrade operation.

View File

@ -23,7 +23,7 @@ one or more letters, as follows:
A slash-separated list of letters is used when the alarm can be triggered with
one of several severity levels.
An asterisk \(\*\) indicates the management-affecting severity, if any. A
An asterisk (\*) indicates the management-affecting severity, if any. A
management-affecting alarm is one that cannot be ignored at the indicated
severity level or higher by using relaxed alarm rules during an orchestrated
patch or upgrade operation.

View File

@ -75,7 +75,7 @@ commands to manage containerized applications provided as part of |prod|.
| updated_at | 2022-06-21T03:13:01.051293+00:00 |
+---------------+----------------------------------+
- Use the following command to upload application Helm chart\(s\) and
- Use the following command to upload application Helm chart\(s) and
manifest.
.. code-block:: none

View File

@ -24,7 +24,7 @@ least the following |SANs|: ``DNS:registry.local``, ``DNS:registry.central``,
IP Address:<oam-floating-ip-address>, IP Address:<mgmt-floating-ip-address>.
Use the :command:`system addrpool-list` command to get the |OAM| floating IP
Address and management floating IP Address for your system. You can add any
additional |DNS| entry\(s\) that you have set up for your |OAM| floating IP
additional |DNS| entry\(s) that you have set up for your |OAM| floating IP
Address.
.. note::
@ -42,7 +42,7 @@ an expired or soon to expire certificate.
.. rubric:: |prereq|
Obtain an intermediate or Root |CA|-signed certificate and key from a trusted
intermediate or Root Certificate Authority \(|CA|\). Refer to the documentation
intermediate or Root Certificate Authority (|CA|). Refer to the documentation
for the external Root |CA| that you are using, on how to create public
certificate and private key pairs, signed by an intermediate or Root |CA|, for
HTTPS.
@ -54,7 +54,7 @@ using openssl <create-certificates-locally-using-openssl>` to create an
Intermediate or test Root |CA| certificate and key, and use it to sign test
certificates.
Put the Privacy Enhanced Mail \(PEM\) encoded versions of the certificate and
Put the Privacy Enhanced Mail (PEM) encoded versions of the certificate and
key in a single file, and copy the file to the controller host.
Also obtain the certificate of the intermediate or Root CA that signed the
@ -87,7 +87,7 @@ information, see, :ref:`Display Certificates Installed on a System <utility-scri
#. Update the Docker registry certificate using the
:command:`certificate-install` command.
Set the mode (``-m`` or ``--mode``) parameter to docker\_registry.
Set the mode (``-m`` or ``--mode``) parameter to docker_registry.
.. code-block:: none

View File

@ -60,7 +60,7 @@ see :ref:`Kubernetes CPU Manager Policies <kubernetes-cpu-manager-policies>`.
When using the static CPU manager policy before increasing the number of
platform CPUs or changing isolated CPUs to application CPUs on a host, ensure
that no pods on the host are making use of any isolated CPUs that will be
affected. Otherwise, the pod\(s\) will transition to a Topology Affinity Error
affected. Otherwise, the pod\(s) will transition to a Topology Affinity Error
state. Although not strictly necessary, the simplest way to do this on systems
other than |AIO-SX| is to administratively lock the host, causing all the
pods to be restarted on an alternate host, before changing CPU assigned
@ -73,9 +73,9 @@ Kubernetes will report a new **windriver.com/isolcpus** resource for each
worker node. This corresponds to the application-isolated CPUs. Pods in the
**Best-effort** or **Burstable** |QoS| class may specify some number of
**windriver.com/isolcpus** resources and the pod will be scheduled on a host
\(and possibly |NUMA| node depending on topology manager policy\) with
\(and possibly |NUMA| node depending on topology manager policy) with
sufficient application-isolated cores available, and the container requesting
the resource will be affined \(and restricted\) to those CPUs via cgroups.
the resource will be affined (and restricted) to those CPUs via cgroups.
Pods in the Guaranteed |QoS| class should not specify **windriver.com/isolcpus**
resources as they will be allocated but not used. If there are multiple

View File

@ -15,9 +15,9 @@ For example:
$ docker login registry.local:9001 -u <keystoneUserName> -p <keystonePassword>
An authorized administrator \('admin' and 'sysinv'\) can perform any Docker
action. Regular users can only interact with their own repositories \(i.e.
registry.local:9001/<keystoneUserName>/\). Any authenticated user can pull from
An authorized administrator ('admin' and 'sysinv') can perform any Docker
action. Regular users can only interact with their own repositories (i.e.
registry.local:9001/<keystoneUserName>/). Any authenticated user can pull from
the following list of public images:
.. _kubernetes-admin-tutorials-authentication-and-authorization-d383e50:

View File

@ -22,8 +22,8 @@ end-users' Kubernetes applications. |prod| recommends to install a Helm v3
client on a remote workstation, so that non-admin (and admin) end-users can
manage their Kubernetes applications remotely.
Upon system installation, local Helm repositories \(containing |prod-long|
packages\) are created and added to the Helm repo list.
Upon system installation, local Helm repositories (containing |prod-long|
packages) are created and added to the Helm repo list.
Use the following command to list these local Helm repositories:
@ -34,8 +34,8 @@ Use the following command to list these local Helm repositories:
starlingx `http://127.0.0.1:8080/helm_charts/starlingx`
stx-platform `http://127.0.0.1:8080/helm_charts/stx-platform`
Where the `stx-platform` repo holds helm charts of StarlingX Applications \(see
next section\) of the |prod| platform itself, while the `starlingx` repo holds
Where the `stx-platform` repo holds helm charts of StarlingX Applications (see
next section) of the |prod| platform itself, while the `starlingx` repo holds
helm charts of optional StarlingX applications, such as Openstack. The admin
user can add charts to these local repos and regenerate the index to use these
charts, and add new remote repositories to the list of known repos.

View File

@ -10,7 +10,7 @@ Use the |prod| system application commands to manage containerized application
deployment from the command-line.
|prod| application management provides a wrapper around FluxCD and Kubernetes
Helm \(see `https://github.com/helm/helm <https://github.com/helm/helm>`__\)
Helm (see `https://github.com/helm/helm <https://github.com/helm/helm>`__)
for managing containerized applications. FluxCD is a tool for managing multiple
Helm charts with dependencies by centralizing all configurations in a single
FluxCD YAML definition and providing life-cycle hooks for all Helm releases.

View File

@ -29,15 +29,15 @@ Static policy customizations
----------------------------
- Pods in the **kube-system** namespace are affined to platform cores
only. Other pod containers \(hosted applications\) are restricted to
only. Other pod containers (hosted applications) are restricted to
running on either the application or isolated cores. CFS quota
throttling for Guaranteed QoS pods is disabled.
- When using the static policy, improved performance can be achieved if
you also use the Isolated CPU behavior as described at :ref:`Isolating CPU Cores to Enhance Application Performance <isolating-cpu-cores-to-enhance-application-performance>`.
- For Kubernetes pods with a **Guaranteed** QoS \(see `https://kubernetes.io/docs/tasks/configure-pod-container/quality-service-pod/ <https://kubernetes.io/docs/tasks/configure-pod-container/quality-service-pod/>`__
for background information\), CFS quota throttling is disabled as it
- For Kubernetes pods with a **Guaranteed** QoS (see `https://kubernetes.io/docs/tasks/configure-pod-container/quality-service-pod/ <https://kubernetes.io/docs/tasks/configure-pod-container/quality-service-pod/>`__
for background information), CFS quota throttling is disabled as it
causes performance degradation.
- Kubernetes pods are prevented by default from running on CPUs with an

View File

@ -56,13 +56,13 @@ The backup contains details as listed below:
All platform configuration data and files required to fully restore the
system to a working state following the platform restore procedure.
- \(Optional\) Any end user container images in **registry.local**; that
- (Optional) Any end user container images in **registry.local**; that
is, any images other than |org| system and application images.
|prod| system and application images are repulled from their
original source, external registries during the restore procedure.
- Home directory 'sysadmin' user, and all |LDAP| user accounts
\(item=/etc\)
(item=/etc)
- Patching and package repositories:

View File

@ -7,8 +7,8 @@
Restore Platform System Data and Storage
========================================
You can perform a system restore \(controllers, workers, including or excluding
storage nodes\) of a |prod| cluster from a previous system backup and bring it
You can perform a system restore (controllers, workers, including or excluding
storage nodes) of a |prod| cluster from a previous system backup and bring it
back to the operational state it was when the backup procedure took place.
.. rubric:: |context|
@ -31,14 +31,14 @@ details on the backup.
the backup was made. You cannot use this backup file to restore the system
to different hardware.
To restore the backup, use the same version of the boot image \(ISO\) that
To restore the backup, use the same version of the boot image (ISO) that
was used at the time of the original installation.
The |prod| restore supports the following optional modes:
.. _restoring-starlingx-system-data-and-storage-ol-tw4-kvc-4jb:
- To keep the Ceph cluster data intact \(false - default option\), use the
- To keep the Ceph cluster data intact (false - default option), use the
following parameter, when passing the extra arguments to the Ansible Restore
playbook command:
@ -46,7 +46,7 @@ The |prod| restore supports the following optional modes:
wipe_ceph_osds=false
- To wipe the Ceph cluster entirely \(true\), where the Ceph cluster will
- To wipe the Ceph cluster entirely (true), where the Ceph cluster will
need to be recreated, or if the Ceph partition was wiped somehow before or
during reinstall, use the following parameter:
@ -57,9 +57,9 @@ The |prod| restore supports the following optional modes:
Restoring a |prod| cluster from a backup file is done by re-installing the
ISO on controller-0, running the Ansible Restore Playbook, applying updates
\(patches\), unlocking controller-0, and then powering on, and unlocking the
\(patches), unlocking controller-0, and then powering on, and unlocking the
remaining hosts, one host at a time, starting with the controllers, and then
the storage hosts, ONLY if required, and lastly the compute \(worker\) hosts.
the storage hosts, ONLY if required, and lastly the compute (worker) hosts.
.. rubric:: |prereq|
@ -92,7 +92,7 @@ conditions are in place:
host manually for network boot immediately after powering it on.
- If you are restoring a |prod-dc| subcloud first, ensure it is in
an **unmanaged** state on the Central Cloud \(SystemController\) by using
an **unmanaged** state on the Central Cloud (SystemController) by using
the following commands:
.. code-block:: none
@ -293,8 +293,8 @@ conditions are in place:
#. If :command:`wipe_ceph_osds` is set to **true**, reinstall the
storage hosts.
#. If :command:`wipe_ceph_osds` is set to **false** \(default
option\), do not reinstall the storage hosts.
#. If :command:`wipe_ceph_osds` is set to **false** (default
option), do not reinstall the storage hosts.
.. caution::
Do not reinstall or power off the storage hosts if you want to
@ -302,7 +302,7 @@ conditions are in place:
will lead to data loss.
#. Ensure that the Ceph cluster is healthy. Verify that the three Ceph
monitors \(controller-0, controller-1, storage-0\) are running in
monitors (controller-0, controller-1, storage-0) are running in
quorum.
.. code-block:: none
@ -332,15 +332,15 @@ conditions are in place:
If the message HEALTH_WARN appears, wait a few minutes and then try
again. If the warning condition persists, consult the public
documentation for troubleshooting Ceph monitors \(for example,
documentation for troubleshooting Ceph monitors (for example,
`http://docs.ceph.com/docs/master/rados/troubleshooting/troubleshootin
g-mon/
<http://docs.ceph.com/docs/master/rados/troubleshooting/troubleshootin
g-mon/>`__\).
g-mon/>`__).
#. Restore the compute \(worker\) hosts, one at a time.
#. Restore the compute (worker) hosts, one at a time.
Restore the compute \(worker\) hosts following the same procedure used to
Restore the compute (worker) hosts following the same procedure used to
restore controller-1.
#. Allow Calico and Coredns pods to be recovered by Kubernetes. They should
@ -396,7 +396,7 @@ conditions are in place:
are not included as part of the backup and restore procedures.
- After restoring a |prod-dc| subcloud, you need to bring it back
to the **managed** state on the Central Cloud \(SystemController\), by
to the **managed** state on the Central Cloud (SystemController), by
using the following commands:
.. code-block:: none

View File

@ -39,7 +39,7 @@ and target it at controller-0.
#. Provide either a customized Ansible hosts file specified using the ``-i``
option, or use the default one in the Ansible configuration directory
\(that is, /etc/ansible/hosts\).
(that is, /etc/ansible/hosts).
#. If using a customized file, change to the ``<br>`` directory created
in the previous step.
@ -106,7 +106,7 @@ and target it at controller-0.
~(keystone_admin)]$ ansible-playbook backup.yml --limit sm5 -i $HOME/br_test/hosts --ask-vault-pass -e "host_backup_dir=$HOME/br_test override_files_dir=$HOME/override_dir"
The generated backup tar file can be found in <host\_backup\_dir>, that
The generated backup tar file can be found in <host_backup_dir>, that
is, /home/sysadmin, by default. You can overwrite it using the **-e**
option on the command line or in an override file.

View File

@ -48,12 +48,12 @@ Other ``-e`` command line options:
- **Optional**: You can select one of the following restore modes:
- To keep the Ceph cluster data intact \(false - default option\), use the
- To keep the Ceph cluster data intact (false - default option), use the
following parameter:
:command:`wipe_ceph_osds=false`
- To wipe the Ceph cluster entirely \(true\), where the Ceph cluster will
- To wipe the Ceph cluster entirely (true), where the Ceph cluster will
need to be recreated, use the following parameter:
:command:`wipe_ceph_osds=true`
@ -133,7 +133,7 @@ Other ``-e`` command line options:
.. rubric:: |postreq|
After running restore\_platform.yml playbook, you can restore the local
After running restore_platform.yml playbook, you can restore the local
registry images.
.. note::

View File

@ -31,7 +31,7 @@ In this method you can run Ansible Restore playbook and point to controller-0.
#. Provide an inventory file, either a customized one that is specified
using the ``-i`` option, or the default one that is in the Ansible
configuration directory \(that is, /etc/ansible/hosts\). You must
configuration directory (that is, /etc/ansible/hosts). You must
specify the floating |OAM| IP of the controller host. For example, if the
host name is |prefix|\_Cluster, the inventory file should have an entry
called |prefix|\_Cluster.
@ -54,12 +54,12 @@ In this method you can run Ansible Restore playbook and point to controller-0.
where ``optional-extra-vars`` can be:
- To keep Ceph data intact \(false - default option\), use the
- To keep Ceph data intact (false - default option), use the
following parameter:
:command:`wipe_ceph_osds=false`
- To start with an empty Ceph cluster \(true\), where the Ceph
- To start with an empty Ceph cluster (true), where the Ceph
cluster will need to be recreated, use the following parameter:
:command:`wipe_ceph_osds=true`
@ -103,7 +103,7 @@ In this method you can run Ansible Restore playbook and point to controller-0.
on controller-0.
- The :command:`ansible_remote_tmp` should be set to a new
directory \(not required to create it ahead of time\) under
directory (not required to create it ahead of time) under
/home/sysadmin on controller-0 using the ``-e`` option on the command
line.

View File

@ -17,12 +17,12 @@ of restoring the underlying platform.
.. note::
Data stored in Ceph such as Glance images, Cinder volumes or volume backups
or Rados objects \(images stored in ceph\) are not backed up automatically.
or Rados objects (images stored in ceph) are not backed up automatically.
.. _back-up-openstack-ul-ohv-x3k-qmb:
- To backup glance images use the image\_backup.sh script. For example:
- To backup glance images use the ``image_backup.sh`` script. For example:
.. code-block:: none

View File

@ -37,8 +37,8 @@ You can restore |prod-os| from a backup with or without Ceph.
- Restore only |prod-os| system data. This option will not restore the
Ceph data \(that is, it will not run commands like :command:`rbd
import`\). This procedure will preserve any existing Ceph data at
Ceph data (that is, it will not run commands like :command:`rbd
import`). This procedure will preserve any existing Ceph data at
restore-time.
- Restore |prod-os| system data, Cinder volumes and Glance images. You'll

View File

@ -59,8 +59,8 @@ pci-passthrough interface.
The name assigned to the data network.
**<type>**
The type of data network to be created \(**flat**, **vlan**, or
**vxlan**\)
The type of data network to be created (**flat**, **vlan**, or
**vxlan**)
.. note::
**vxlan** is only applicable to |prod-os|.
@ -86,7 +86,7 @@ pci-passthrough interface.
For the |prod-os| application, after creating a data network of the VLAN or
VXLAN type, you can assign one or more segmentation ranges consisting of a set
of consecutive VLAN IDs \(for VLANs\) or VNIs \(for VXLANs\) using the
of consecutive VLAN IDs (for VLANs) or VNIs (for VXLANs) using the
:command:`openstack network segment range create` command. Segmentation ranges
are required in order to set up project networks.

View File

@ -13,4 +13,4 @@ The command for performing the mapping has the format:
.. code-block:: none
~(keystone_admin)]$ system interfacedatanetworkassign <host\_name> <interface\_uuid> <datanetwork\_uuid>
~(keystone_admin)]$ system interfacedatanetworkassign <host_name> <interface_uuid> <datanetwork_uuid>

View File

@ -13,7 +13,7 @@ A data network represents a Layer 2 physical or virtual network, or set of
virtual networks, used to provide the underlying network connectivity needed
to support the application networks. Multiple data networks may be configured
as required, and realized over the same or different physical networks. Access
to external networks is typically \(although not necessarily\) granted to
to external networks is typically (although not necessarily) granted to
worker nodes using a data network. The extent of this connectivity, including
access to the open internet, is application dependent.

View File

@ -84,6 +84,6 @@ Labels for Network Connections
------------------------------
Network connections in the topology window may be labeled with the data
interface name \(displayed above the connection line\) and LLDP neighbor
information \(displayed below the connection line\). You can show or hide the
labels using a button above the lists \(**Show Labels** or **Hide Labels**\).
interface name (displayed above the connection line) and LLDP neighbor
information (displayed below the connection line). You can show or hide the
labels using a button above the lists (**Show Labels** or **Hide Labels**).

View File

@ -58,8 +58,8 @@ To make interface changes, you must lock the compute host first.
.. image:: /shared/figures/datanet/jow1442607685238.png
#. Enter the IPv4 or IPv6 address and netmask \(for example,
192.168.1.3/24\), and then click **Create Address**.
#. Enter the IPv4 or IPv6 address and netmask (for example,
192.168.1.3/24), and then click **Create Address**.
The new address is added to the **Address List**.

View File

@ -39,7 +39,7 @@ where
is the default gateway
**metric**
is the cost of the route \(the number of hops\)
is the cost of the route (the number of hops)
To delete routes, use the following command.

View File

@ -64,7 +64,7 @@ You can use the CLI to add segmentation ranges to data networks.
is the data network associated with the range.
**network type**
is the network type \(VLAN/VXLAN\) of the range.
is the network type (VLAN/VXLAN) of the range.
**minimum**
is the minimum value of the segmentation range.

View File

@ -28,7 +28,7 @@ where:
**<interface name>**
is the name of the interface
**<mtu\_size>**
**<mtu_size>**
is the new |MTU| value
For example:

View File

@ -7,7 +7,7 @@ Configure Data Interfaces for VXLANs
====================================
For data interfaces attached to VXLAN-based data networks, endpoint IP
addresses, static or dynamic from a IP Address pool\) and possibly IP Routes
addresses, static or dynamic from a IP Address pool) and possibly IP Routes
are additionally required on the host data interfaces.
See :ref:`VXLAN Data Network Setup Completion

View File

@ -100,14 +100,14 @@ For each of the above procedures, configure the node interface specifying the
The name or |UUID| of the Ethernet interface to use.
**ip4\_mode**
The mode for assigning IPv4 addresses to a data interface \(static or
pool.\)
The mode for assigning IPv4 addresses to a data interface (static or
pool.)
**ip6\_mode**
The mode for assigning IPv6 addresses to a data interface \(static or
pool.\)
The mode for assigning IPv6 addresses to a data interface (static or
pool.)
**addr\_pool**
**addr_pool**
The name of an IPv4 or IPv6 address pool, for use with the pool mode
of IP address assignment for data interfaces.

View File

@ -6,16 +6,16 @@
Dynamic VXLAN
=============
|prod-os| supports dynamic mode \(learning\) VXLAN implementation that has each
|prod-os| supports dynamic mode (learning) VXLAN implementation that has each
vSwitch instance registered on the network for a particular IP multicast group,
|MAC| addresses, and |VTEP| endpoints that are populated based on neutron
configuration data.
The IP multicast group, \(for example, 239.1.1.1\), is input when a new
The IP multicast group, (for example, 239.1.1.1), is input when a new
neutron data network is provisioned. The selection of the IP multicast group
constraints flooding to only those nodes that have registered for the specified
group. The IP multicast network can work in both a single subnet \(that is,
local Layer2 environment\) or can span Layer3 segments in the customer network
group. The IP multicast network can work in both a single subnet (that is,
local Layer2 environment) or can span Layer3 segments in the customer network
for more complex routing requirements but requires IP multicast enabled routers.
.. only:: starlingx
@ -70,7 +70,7 @@ segmentation ranges using the |CLI|.
#. Create a VXLAN data network, see :ref:`Adding Data Networks
<adding-data-networks-using-the-cli>`.
#. Add segmentation ranges to dynamic |VXLAN| \(Multicast |VXLAN|\) data
#. Add segmentation ranges to dynamic |VXLAN| (Multicast |VXLAN|) data
networks, see :ref:`Adding Segmentation Ranges Using the CLI
<adding-segmentation-ranges-using-the-cli>`.

View File

@ -59,7 +59,7 @@ To make interface changes, you must lock the compute node first.
**ifname**
is the name of the interface
**ip\_address**
**ip_address**
is an IPv4 or IPv6 address
**prefix**

View File

@ -39,21 +39,21 @@ where:
is a name used to select the pool during data interface setup
**<network>**
is the subnet and mask for the range \(for example, **192.168.1.0**\)
is the subnet and mask for the range (for example, **192.168.1.0**)
**<prefix>**
is the subnet mask, expressed in network prefix length notation \(for
example, **24**\)
is the subnet mask, expressed in network prefix length notation (for
example, **24**)
**<assign\_order>**
is the order in which to assign addresses from the pool \(random or
sequential\). The default is random.
**<assign_order>**
is the order in which to assign addresses from the pool (random or
sequential). The default is random.
**<addr\_ranges>**
**<addr_ranges>**
is a set of IP address ranges to use for assignment, where the start
and end IP address of each range is separated by a dash, and the ranges
are separated by commas \(for example, **192.168.1.10-192.168.1.20,
192.168.1.35-192.168.1.45**\). If no range is specified, the full range is
are separated by commas (for example, **192.168.1.10-192.168.1.20,
192.168.1.35-192.168.1.45**). If no range is specified, the full range is
used.
.. _managing-ip-address-pools-using-the-cli-section-N10109-N1001F-N10001:

View File

@ -51,7 +51,7 @@ To make interface changes, you must lock the compute node first.
A name used for selecting the pool during data interface setup.
**Network Address**
The subnet for the range \(for example, **192.168.1.0/24**\).
The subnet for the range (for example, **192.168.1.0/24**).
**Allocation Order**
The order for assigning addresses. You can select **Sequential** or
@ -59,8 +59,8 @@ To make interface changes, you must lock the compute node first.
**Address Range**
One or more ranges, where the start and end IP address of each range
is separated by a dash, and the ranges are separated by commas \(for
example, **192.168.1.10-192.168.1.20, 192.168.1.35-192.168.1.45**\).
is separated by a dash, and the ranges are separated by commas (for
example, **192.168.1.10-192.168.1.20, 192.168.1.35-192.168.1.45**).
If no range is specified, the full range is used.
.. rubric:: |postreq|

View File

@ -6,11 +6,11 @@
VXLAN Data Networks
===================
Virtual eXtensible Local Area Networks \(|VXLANs|\) data networks are an
Virtual eXtensible Local Area Networks (|VXLANs|) data networks are an
alternative to |VLAN| data networks.
A |VXLAN| data network is implemented over a range of |VXLAN| Network
Identifiers \(|VNIs|.\) This is similar to the |VLAN| option, but allows
Identifiers (|VNIs|.) This is similar to the |VLAN| option, but allows
multiple data networks to be defined over the same physical network using
unique |VNIs| defined in segmentation ranges.

View File

@ -9,12 +9,12 @@ Deployment Configurations
A variety of |prod-long| deployment configuration options are supported.
**All-in-one Simplex**
A single physical server providing all three cloud functions \(controller,
worker and storage\).
A single physical server providing all three cloud functions (controller,
worker and storage).
**All-in-one Duplex \(up to 50 worker nodes\)**
**All-in-one Duplex (up to 50 worker nodes)**
Two HA-protected physical servers, both running all three cloud functions
\(controller, worker and storage\), optionally with up to 50 worker nodes
(controller, worker and storage), optionally with up to 50 worker nodes
added to the cluster.
**Standard with Storage Cluster on Controller Nodes**

View File

@ -23,7 +23,7 @@ A number of components are common to most |prod| deployment configurations.
For standard with controller storage deployment configurations, the
controller nodes/functions run a small-scale Ceph cluster using one or more
disks \(|SATA|, |SAS|, |SSD| and/or |NVMe|\) as the ceph |OSDs|. This
disks (|SATA|, |SAS|, |SSD| and/or |NVMe|) as the ceph |OSDs|. This
cluster provides the storage backend for Kubernetes' |PVCs|.
In most configurations, the controller nodes/functions are part of a two
@ -35,8 +35,8 @@ A number of components are common to most |prod| deployment configurations.
**Storage Node / Function**
For Standard with Dedicated Storage deployment configurations, the storage
nodes run a large scale Ceph cluster using disks \(|SATA|, |SAS|, |SSD| and
/or |NVMe|\) across 2-9 storage nodes as Ceph |OSDs|. This provides the
nodes run a large scale Ceph cluster using disks (|SATA|, |SAS|, |SSD| and
/or |NVMe|) across 2-9 storage nodes as Ceph |OSDs|. This provides the
storage backend for Kubernetes' |PVCs|.
In most configurations the storage nodes/functions are part of a HA
@ -50,20 +50,20 @@ A number of components are common to most |prod| deployment configurations.
**L2 Switches and L2 Networks**
A single physical switch may support multiple L2 networks.
**Operations, Administration and Management (OAM) Network \(Controller Nodes Only\)**
**Operations, Administration and Management (OAM) Network (Controller Nodes Only)**
The network on which all external StarlingX platform APIs are exposed,
including platform REST APIs \(Keystone, StarlingX, Kubernetes\), the
including platform REST APIs (Keystone, StarlingX, Kubernetes), the
Horizon Web interface, |SSH| and |SNMP|.
This is typically a 1GE network.
**Management Network \(All Nodes\)**
A private network \(i.e. not connected externally\) used for internal
**Management Network (All Nodes)**
A private network (i.e. not connected externally) used for internal
StarlingX monitoring and control, and container access to storage cluster.
This is typically a 10GE network.
**Cluster Host Network \(All Nodes\)**
**Cluster Host Network (All Nodes)**
The cluster host network is used for Kubernetes management and control, as
well as private container networking. The |CNI| service, Calico, provides
private tunneled networking between hosted containers on the cluster host
@ -83,12 +83,12 @@ A number of components are common to most |prod| deployment configurations.
Containers' network endpoints can be exposed externally with 'NodePort'
Kubernetes services, exposing selected application containers' network
ports on *all* interfaces \(e.g. external cluster host interfaces\) of
ports on *all* interfaces (e.g. external cluster host interfaces) of
both controller nodes and *all* worker nodes. This would typically be
done either directly to the application containers service or through
an ingress controller service. HA would be achieved through either an
external HA load balancer across two or more worker nodes or simply
using multiple records \(two or more destination worker node IPs\) for
using multiple records (two or more destination worker node IPs) for
the application's external DNS Entry.
Containers' network endpoints can also be exposed through |BGP| within
@ -111,13 +111,13 @@ A number of components are common to most |prod| deployment configurations.
nodes. This is typically done either directly to the application
containers service or through an ingress controller service. HA can be
achieved through either an external HA load balancer across two or more
worker nodes or simply using multiple records \(two or more destination
worker node IP addresses\) for the application's external DNS Entry.
worker nodes or simply using multiple records (two or more destination
worker node IP addresses) for the application's external DNS Entry.
The use of Container Networking Calico |BGP| to advertise containers'
network endpoints is not available in this scenario.
**Additional External Network\(s\) or Data Networks \(Worker & AIO Nodes Only\)**
**Additional External Network\(s) or Data Networks (Worker & AIO Nodes Only)**
Networks on which ingress controllers and/or hosted application containers
expose their Kubernetes service, for example, through a NodePort service.
Node interfaces to these networks are configured as platform class
@ -128,13 +128,13 @@ A number of components are common to most |prod| deployment configurations.
hosted application containers to have interfaces directly connected to the
host's interface via pci-passthru or |SRIOV|.
**IPMI Network \(All Nodes\)**
**IPMI Network (All Nodes)**
An optional network on which |IPMI| interfaces of all nodes are connected.
The |IPMI| network must be L3/IP reachable from the controller's |OAM|
interfaces.
**PxeBoot Network \(All Nodes\)**
**PxeBoot Network (All Nodes)**
An *optional* network over which nodes net boot from controllers.
By default, controllers network boot other nodes over the management

View File

@ -61,7 +61,7 @@ management and cluster host network.
|org| recommends a 10GE shared management and cluster host network with
|LAG| for direct connections. If the management
network must be 1GE \(to support PXE booting\), then a separate 10GE
network must be 1GE (to support PXE booting), then a separate 10GE
cluster host network with |LAG| is also
recommended. The use of |LAG| addresses failover
considerations unique to peer-to-peer connections.
@ -81,8 +81,8 @@ provide support for small scale deployments on the Intel Xeon D family of
processors using a smaller memory and CPU footprint than the standard Simplex
configuration.
For low-cost or low-power applications with minimal performance demands \(40
containers or fewer\), |prod| Simplex can be deployed on a server with a
For low-cost or low-power applications with minimal performance demands (40
containers or fewer), |prod| Simplex can be deployed on a server with a
single Intel Xeon D class processor. The platform-reserved memory and the
maximum number of worker threads are reduced by default, but can be
reconfigured if required.

View File

@ -63,8 +63,8 @@ provide support for small scale deployments on the Intel Xeon D family of
processors using a smaller memory and CPU footprint than the standard Simplex
configuration.
For low-cost or low-power applications with minimal performance demands \(40
Containers or fewer\), |prod| Simplex can be deployed on a server with a
For low-cost or low-power applications with minimal performance demands (40
Containers or fewer), |prod| Simplex can be deployed on a server with a
single Intel Xeon D class processor. The platform-reserved memory and the
maximum number of worker threads are reduced by default, but can be
reconfigured as required.

View File

@ -9,12 +9,12 @@ Deployment Options
A variety of |prod-long| deployment configuration options are supported.
**All-in-one Simplex**
A single physical server providing all three cloud functions \(controller,
worker and storage\).
A single physical server providing all three cloud functions (controller,
worker and storage).
**All-in-one Duplex \(up to 50 worker nodes\)**
**All-in-one Duplex (up to 50 worker nodes)**
Two HA-protected physical servers, both running all three cloud functions
\(controller, worker and storage\), optionally with up to 50 worker nodes
(controller, worker and storage), optionally with up to 50 worker nodes
added to the cluster.
**Standard with Storage Cluster on Controller Nodes**
@ -29,6 +29,6 @@ A variety of |prod-long| deployment configuration options are supported.
information, see the :ref:`Storage
<storage-configuration-storage-resources>` guide.
All |prod| systems can use worker platforms \(worker hosts, or the worker
function on a simplex or duplex system\) configured for either standard or
All |prod| systems can use worker platforms (worker hosts, or the worker
function on a simplex or duplex system) configured for either standard or
low-latency worker function performance profiles.

View File

@ -7,7 +7,7 @@ Standard Configuration with Dedicated Storage
=============================================
Deployment of |prod| with dedicated storage nodes provides the highest capacity
\(single region\), performance, and scalability.
\(single region), performance, and scalability.
.. image:: /shared/figures/deploy_install_guides/starlingx-deployment-options-dedicated-storage.png
:width: 800
@ -28,9 +28,9 @@ storage, and network interfaces can be scaled to meet requirements.
Storage nodes provide a large scale Ceph cluster for the storage backend for
Kubernetes |PVCs|. They are deployed in replication groups of either two or
three for redundancy. For a system configured to use two storage hosts per
replication group, a maximum of eight storage hosts \(four replication groups\)
replication group, a maximum of eight storage hosts (four replication groups)
are supported. For a system with three storage hosts per replication group, up
to nine storage hosts \(three replication groups\) are supported. The system
to nine storage hosts (three replication groups) are supported. The system
provides redundancy and scalability through the number of Ceph |OSDs| installed
in a storage node group, with more |OSDs| providing more capacity and better
storage performance. The scalability and performance of the storage function is

View File

@ -26,7 +26,7 @@ You can add an arbitrary number of hosts using a single CLI command.
~[keystone_admin]$ system host-bulk-add <xml_file>
where <xml\_file> is the name of the prepared XML file.
where <xml_file> is the name of the prepared XML file.
#. Power on the hosts to be added, if required.

View File

@ -81,7 +81,7 @@ scripting an initial setup.
- storage
**<subfunctions>**
are the host personality subfunctions \(used only for a worker host\).
are the host personality subfunctions (used only for a worker host).
For a worker host, the only valid value is worker,lowlatency to enable
a low-latency performance profile. For a standard performance profile,
@ -95,53 +95,53 @@ scripting an initial setup.
is a string describing the location of the host
**<console>**
is the output device to use for message display on the host \(for
example, tty0\). The default is ttys0, 115200.
is the output device to use for message display on the host (for
example, tty0). The default is ttys0, 115200.
**<install\_output>**
is the format for console output on the host \(text or graphical\). The
**<install_output>**
is the format for console output on the host (text or graphical). The
default is text.
.. note::
The graphical option currently has no effect. Text-based
installation is used regardless of this setting.
**<boot\_device>**
**<boot_device>**
is the host device for boot partition, relative to /dev. The default is
sda.
**<rootfs\_device>**
**<rootfs_device>**
is a logical volume cgts-vg/root-lv. The default is sda, it should be
the same value as specified for the boot_device.
**<mgmt\_mac>**
**<mgmt_mac>**
is the |MAC| address of the port connected to the internal management
or |PXE| boot network.
**<mgmt\_ip>**
**<mgmt_ip>**
is the IP address of the port connected to the internal management or
|PXE| boot network, if static IP address allocation is used.
.. note::
The <mgmt\_ip> option is not used for a controller node.
The <mgmt_ip> option is not used for a controller node.
**<ttys\_dcd>**
**<ttys_dcd>**
is set to **True** to have any active console session automatically
logged out when the serial console cable is disconnected, or **False**
to disable this behavior. The server must support data carrier detect
on the serial console port.
**<bm\_type>**
**<bm_type>**
is the board management controller type. Use bmc.
**<bm\_ip>**
is the board management controller IP address \(used for external
access to board management controllers over the |OAM| network\)
**<bm_ip>**
is the board management controller IP address (used for external
access to board management controllers over the |OAM| network)
**<bm\_username>**
**<bm_username>**
is the username for board management controller access
**<bm\_password>**
**<bm_password>**
is the password for board management controller access
For example:

View File

@ -34,30 +34,30 @@ valid values, refer to the CLI documentation.
+-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| subfunctions | For a worker host, an optional element to enable a low-latency performance profile. |
+-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| mgmt\_mac | The MAC address of the management interface. |
| mgmt_mac | The MAC address of the management interface. |
+-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| mgmt\_ip | The IP address of the management interface. |
| mgmt_ip | The IP address of the management interface. |
+-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| bm\_ip | The IP address of the board management controller. |
| bm_ip | The IP address of the board management controller. |
+-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| bm\_type | The board management controller type. |
| bm_type | The board management controller type. |
+-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| bm\_username | The username for board management controller authentication. |
| bm_username | The username for board management controller authentication. |
+-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| bm\_password | The password for board management controller authentication. |
| bm_password | The password for board management controller authentication. |
+-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| power\_on | An empty element. If present, powers on the host automatically using the specified board management controller. |
| power_on | An empty element. If present, powers on the host automatically using the specified board management controller. |
+-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| install\_output | The display mode to use during installation \(text or graphical\). The default is **text**. |
| install_output | The display mode to use during installation (text or graphical). The default is **text**. |
| | |
| | .. note:: |
| | The graphical option currently has no effect. Text-based installation is used regardless of this setting. |
+-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| console | If present, this element specifies the port, and if applicable the baud, for displaying messages. If the element is empty or not present, the default setting **ttyS0,115200** is used. |
+-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| rootfs\_device | The root filesystem is now a logical volume cgts-vg/root-lv. This value when shown should be the same value as the boot_device. |
| rootfs_device | The root filesystem is now a logical volume cgts-vg/root-lv. This value when shown should be the same value as the boot_device. |
+-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| boot\_device | The device to use for the boot partition, relative to /dev. |
| boot_device | The device to use for the boot partition, relative to /dev. |
+-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| location | A description of the host location. |
+-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+

View File

@ -16,8 +16,8 @@ host-bulk-export` command, and then use this file with the :command:`system
host-bulk-add` command to re-create the system. If required, you can modify the
file before using it.
The configuration settings \(management |MAC| address, BM IP address, and so
on\) for all nodes except **controller-0** are written to the file.
The configuration settings (management |MAC| address, BM IP address, and so
on) for all nodes except **controller-0** are written to the file.
.. note::
To ensure that the hosts are not powered on unexpectedly, the **power-on**

View File

@ -17,7 +17,7 @@ For a summary of changes that require system or host reinstallation, see
<configuration-changes-requiring-re-installation>`.
To reinstall an entire system, refer to the Installation Guide for your system
type \(for example, Standard or All-in-one\).
type (for example, Standard or All-in-one).
.. note::
To simplify system reinstallation, you can export and reuse an existing
@ -29,8 +29,8 @@ To reinstall the software on a host using the Host Inventory controls, see
|node-doc|: :ref:`Host Inventory <hosts-tab>`. In some cases, you must delete
the host instead, and then re-add it using the standard host installation
procedure. This applies if the system inventory record must be corrected to
complete the configuration change \(for example, if the |MAC| address of the
management interface has changed\).
complete the configuration change (for example, if the |MAC| address of the
management interface has changed).
- :ref:`Reinstalling a System Using an Exported Host Configuration File
<reinstalling-a-system-using-an-exported-host-configuration-file-r7>`

View File

@ -157,7 +157,7 @@ Example build command:
--stream ${BUILD_STREAM}
| This will produce a wheels tarball in your workspace:
| ${MY\_WORKSPACE}/std/build-wheels-${OS}-${BUILD\_STREAM}/stx-${OS}-${BUILD\_STREAM}-wheels.tar
| ${MY_WORKSPACE}/std/build-wheels-${OS}-${BUILD_STREAM}/stx-${OS}-${BUILD_STREAM}-wheels.tar
****************
StarlingX wheels
@ -168,7 +168,7 @@ the build. For CentOs, this means updating the package rpm specfile to
build the wheel and package it in a -wheels package. The names of the
wheels packages to be included in the tarball are listed in the
wheels.inc files in the corresponding repo (ie.
centos\_stable\_wheels.inc).
centos_stable_wheels.inc).
---------------
Building images
@ -178,8 +178,8 @@ The StarlingX Docker images are built using a set of image directives
files, with the base image and wheels tarball as input. The images are
built by the build-stx-images.sh tool, in
stx-root/build-tools/build-docker-images. The build-stx-images.sh tool
will search the StarlingX repos for a corresponding docker\_images.inc
file (ie. centos\_dev\_docker\_images.inc) which contains a list of
will search the StarlingX repos for a corresponding docker_images.inc
file (ie. centos_dev_docker_images.inc) which contains a list of
subdirectories that contain the associated image directives files, which
are processed and built.
@ -286,15 +286,15 @@ Options supported by BUILDER=docker image directives files include:
* LABEL: the image name
* PROJECT: main project name
* DOCKER\_REPO: main project source git repo
* DOCKER\_REF: git branch or tag for main project source repo (default "master")
* DOCKER\_PATCHES: list of patch files to apply to DOCKER\_REPO, relative to the local dir
* DOCKER\_CONTEXT: path to build context source, relative to the local dir (default "docker")
* DOCKER\_FILE: path to Dockerfile, relative to the local dir (default "docker/Dockerfile")
* DOCKER_REPO: main project source git repo
* DOCKER_REF: git branch or tag for main project source repo (default "master")
* DOCKER_PATCHES: list of patch files to apply to DOCKER_REPO, relative to the local dir
* DOCKER_CONTEXT: path to build context source, relative to the local dir (default "docker")
* DOCKER_FILE: path to Dockerfile, relative to the local dir (default "docker/Dockerfile")
.. note::
DOCKER\_CONTEXT and DOCKER\_FILE are mutually exclusive to DOCKER\_REPO, DOCKER\_REF and DOCKER\_PATCHES.
DOCKER_CONTEXT and DOCKER_FILE are mutually exclusive to DOCKER_REPO, DOCKER_REF and DOCKER_PATCHES.
For an example of a BUILDER=docker image, see
https://opendev.org/starlingx/oidc-auth-armada-app/src/branch/master/dex/centos/dex.stable_docker_image
@ -317,11 +317,11 @@ loci include:
* LABEL: the image name
* PROJECT: main project name
* PROJECT\_REPO: main project source git repo
* PROJECT\_REF: git branch or tag for main project source repo
* PIP\_PACKAGES: list of python modules to be installed, beyond those
* PROJECT_REPO: main project source git repo
* PROJECT_REF: git branch or tag for main project source repo
* PIP_PACKAGES: list of python modules to be installed, beyond those
specified by project dependencies or requirements
* DIST\_PACKAGES: additional packages to be installed (eg. RPMs from
* DIST_PACKAGES: additional packages to be installed (eg. RPMs from
repo, configured by base image)
* PROFILES: bindep profiles supported by project to be installed (eg.
apache)
@ -330,7 +330,7 @@ In addition, you can specify a bash command in the CUSTOMIZATION option,
in order to do a modification on the loci-built image.
Example:
stx-upstream/openstack/python-nova/centos/stx-nova.dev\_docker\_image
stx-upstream/openstack/python-nova/centos/stx-nova.dev_docker_image
::
@ -347,7 +347,7 @@ In a case where the image is built without a main project source git
repo, where the main project source is just coming from a wheel, you can
set PROJECT to infra, and loci skips the git clone steps. For example,
stx-nova-api-proxy:
stx-nfv/nova-api-proxy/centos/stx-nova-api-proxy.dev\_docker\_image
stx-nfv/nova-api-proxy/centos/stx-nova-api-proxy.dev_docker_image
::
@ -466,7 +466,7 @@ of the entire image. The tool allows for updates via:
Specifying Python module source
*******************************
The --module-src command-line option (or MODULE\_SRC in an update
The --module-src command-line option (or MODULE_SRC in an update
directives file) allows a designer to specify python module source from
either a directory or git repository. If specifying a git repository,
you can also specify a branch or tag to be fetched, as well as
@ -483,9 +483,9 @@ Customization script
You can optionally provide a customization script to make changes to the
image that cannot be handled by updating software, using the --customize
command-line option (or CUSTOMIZATION\_SCRIPT in an update directives
command-line option (or CUSTOMIZATION_SCRIPT in an update directives
file). You can also provide supporting files with the --extra
command-line option (or EXTRA\_FILES in an update directives file),
command-line option (or EXTRA_FILES in an update directives file),
which will be accessible to the customization script in the
/image-update/extras directory within the update container.

View File

@ -36,7 +36,7 @@ manages the following certificates:
.. certificate-management-for-admin-rest--api-endpoints-ul-zdc-pmk-xnb:
- **DC-AdminEp-Root-CA certificate**: This certificate expires in 1825 days
\(approximately 5 years\). Renewal of this certificate starts 30 days prior
(approximately 5 years). Renewal of this certificate starts 30 days prior
to expiry.
The Root |CA| certificate is renewed on the System Controller. When the

View File

@ -46,8 +46,8 @@ Ensure that all subclouds are managed and online.
.. note::
In a subcloud, if the |CLI| command returns an authentication error
after you source the script /etc/platform/openrc, you can verify
the password on the subcloud by using the :command:`env \| grep OS\_PASSWORD`
after you source the script ``/etc/platform/openrc``, you can verify
the password on the subcloud by using the :command:`env \| grep OS_PASSWORD`
command . If it returns the old password, you will need to run the
:command:`keyring set CGCS admin` command and provide the new admin
password.

View File

@ -22,8 +22,8 @@ The following settings are applied by default:
- alarm restriction type: relaxed
- default instance action: migrate \(This parameter is only applicable to
hosted application |VMs| with the |prefix|-openstack application.\)
- default instance action: migrate (This parameter is only applicable to
hosted application |VMs| with the |prefix|-openstack application.)
To update the default values, use the :command:`dcmanager strategy-config
@ -112,7 +112,7 @@ individual subclouds.
migrate or stop-start — determines whether hosted application |VMs| are
migrated or stopped and restarted when a worker host is upgraded
**subcloud\_name**
**subcloud_name**
The name of the subcloud to use the custom strategy. If this omitted,
the default upgrade strategy is updated.

View File

@ -148,12 +148,12 @@ controller for access by subclouds. For example:
**--max-parallel-subclouds**
Sets the maximum number of subclouds that can be upgraded in parallel
\(default 20\). If this is not specified using the CLI, the values for
(default 20). If this is not specified using the CLI, the values for
max_parallel_subclouds defined for each subcloud group will be used by
default.
**--stop-on-failure**
**false** \(default\) or **true** — determines whether upgrade
**false** (default) or **true** — determines whether upgrade
orchestration failure for a subcloud prevents application to subsequent
subclouds.
@ -257,7 +257,7 @@ controller for access by subclouds. For example:
.. note::
After the *Kubernetes Version Upgrade Distributed Cloud Orchestration
Strategy* has been applied \(or aborted\) it must be deleted before
Strategy* has been applied (or aborted) it must be deleted before
another Kubernetes Version Upgrade Distributed Cloud Orchestration
strategy can be created. If a Kubernetes upgrade strategy application
fails, you must address the issue that caused the failure, then delete

View File

@ -20,7 +20,7 @@ If the Subclouds are in a **Managed** state and if the patching sync status is
.. rubric:: |context|
Only one update strategy can exist at a time. The strategy controls how the
subclouds are updated \(for example, serially or in parallel\).
subclouds are updated (for example, serially or in parallel).
To determine how the nodes on the Central Cloud's RegionOne and each subcloud
are updated, the update strategy refers to separate configuration settings

View File

@ -6,8 +6,8 @@
Customize the Update Configuration for Distributed Cloud Update Orchestration
=============================================================================
You can adjust how the nodes in each system \(Central Cloud's RegionOne and/or
Subclouds\) are updated.
You can adjust how the nodes in each system (Central Cloud's RegionOne and/or
Subclouds) are updated.
.. rubric:: |context|

View File

@ -161,14 +161,14 @@ device image updates, including |FPGA| updates.
**max-parallel-subclouds**
Sets the maximum number of subclouds that can be updated in parallel
\(default 20\).
(default 20).
If this is not specified using the |CLI|, the values for
:command:`max_parallel_subclouds` defined for each subcloud group
will be used by default.
**stop-on-failure**
true or false \(default\) — determines whether update orchestration
true or false (default) — determines whether update orchestration
failure for a subcloud prevents application to subsequent subclouds.
**group**

View File

@ -53,8 +53,8 @@ if using the CLI.
.. note::
Services in a Subcloud authenticate against their local Identity
Provider only \(i.e. Keystone for StarlingX and Kubernetes Service
Accounts for Kubernetes\). This allows the subcloud to not only be
Provider only (i.e. Keystone for StarlingX and Kubernetes Service
Accounts for Kubernetes). This allows the subcloud to not only be
autonomous in the face of disruptions with the Central Region, but also
allows the subcloud to improve service performance since authentication
is localized within the subcloud.
@ -64,7 +64,7 @@ if using the CLI.
Each subcloud can be in a Managed or Unmanaged state.
**Managed**
When a subcloud is in the Managed state, it is updated \(synchronized\)
When a subcloud is in the Managed state, it is updated (synchronized)
immediately with configuration changes made at the System Controller.
This is the normal operating state. Updates may be delayed slightly
depending on network conditions.

View File

@ -42,7 +42,7 @@ following conditions:
- The subclouds must use the Redfish platform management service if it is
an |AIO-SX| subcloud.
- Duplex \(|AIO-DX|/Standard\) upgrades are supported, and they do not
- Duplex (|AIO-DX|/Standard) upgrades are supported, and they do not
require remote install using Redfish.
- Redfish |BMC| is required for orchestrated subcloud upgrades. The install
@ -62,8 +62,8 @@ following conditions:
:ref:`Installing a Subcloud Using Redfish Platform Management Service
<installing-a-subcloud-using-redfish-platform-management-service>`.
- All subclouds are clear of management-affecting alarms \(with the exception of the alarm upgrade
in progress\).
- All subclouds are clear of management-affecting alarms (with the exception of the alarm upgrade
in progress).
- All hosts of all subclouds must be unlocked, enabled, and available.
@ -104,7 +104,7 @@ dcmanager CLI or the Horizon web interface. If you prefer to use Horizon, see
After the System Controller upgrade is completed, wait for 10 minutes for
the **load_sync_status** of all subclouds to be updated.
To identify which subclouds are upgrade-current \(in-sync\), use the
To identify which subclouds are upgrade-current (in-sync), use the
:command:`subcloud list` command. For example:
.. code-block:: none
@ -184,14 +184,14 @@ dcmanager CLI or the Horizon web interface. If you prefer to use Horizon, see
**max-parallel-subclouds**
Sets the maximum number of subclouds that can be upgraded in parallel
\(default 20\).
(default 20).
If this is not specified using the CLI, the values for
:command:`max_parallel_subclouds` defined for each subcloud group
will be used by default.
**stop-on-failure**
**true**\(default\) or **false**— determines whether upgrade
**true**\(default) or **false**— determines whether upgrade
orchestration failure for a subcloud prevents application to subsequent
subclouds.

View File

@ -7,8 +7,8 @@
Install a Subcloud Using Redfish Platform Management Service
============================================================
For subclouds with servers that support Redfish Virtual Media Service \(version
1.2 or higher\), you can use the Central Cloud's CLI to install the ISO and
For subclouds with servers that support Redfish Virtual Media Service (version
1.2 or higher), you can use the Central Cloud's CLI to install the ISO and
bootstrap the subclouds from the Central Cloud.
@ -47,9 +47,9 @@ subcloud, the subcloud installation has these phases:
:command:`load-import` command to allow the import into the
System Controller ``/opt/dc-vault/loads``. The purpose of this is to allow
Redfish install of subclouds referencing a single full copy of the
``bootimage.iso`` at ``/opt/dc-vault/loads``. \(Previously, the full
``bootimage.iso`` at ``/opt/dc-vault/loads``. (Previously, the full
``bootimage.iso`` was duplicated for each :command:`subcloud add`
command\).
command).
.. note::

View File

@ -56,7 +56,7 @@ subcloud, the subcloud installation process has two phases:
- In order to be able to deploy subclouds from either controller, all local
files that are referenced in the ``bootstrap.yml`` file must exist on both
controllers \(for example, ``/home/sysadmin/docker-registry-ca-cert.pem``\).
controllers (for example, ``/home/sysadmin/docker-registry-ca-cert.pem``).
.. rubric:: |proc|
@ -74,8 +74,8 @@ subcloud, the subcloud installation process has two phases:
:start-after: begin-ref-1
:end-before: end-ref-1
#. Update the ISO image to modify installation boot parameters \(if
required\), automatically select boot menu options and add a kickstart file
#. Update the ISO image to modify installation boot parameters (if
required), automatically select boot menu options and add a kickstart file
to automatically perform configurations such as configuring the initial IP
Interface for bootstrapping.
@ -258,7 +258,7 @@ subcloud, the subcloud installation process has two phases:
deployment by monitoring the following log files on the active controller
in the Central Cloud.
/var/log/dcmanager/ansible/<subcloud\_name>\_playbook.output.log
/var/log/dcmanager/ansible/<subcloud_name>\_playbook.output.log
For example:

View File

@ -15,8 +15,8 @@ Platform Management Service.
.. note::
Each subcloud must be on a separate management subnet \(different from the
System Controller and from any other subclouds\).
Each subcloud must be on a separate management subnet (different from the
System Controller and from any other subclouds).
.. _installing-and-provisioning-a-subcloud-section-orn-jkf-t4b:

View File

@ -18,17 +18,17 @@ a different subcloud group, if required. To create a subcloud group, see,
For example, while creating a strategy, if several subclouds can be upgraded or
updated in parallel, they can be grouped together in a subcloud group that
supports parallel upgrades or updates. In this case, the
:command:`max\_parallel\_subclouds`, and :command:`subcloud\_apply\_type` are
:command:`max_parallel_subclouds`, and :command:`subcloud_apply_type` are
**not** specified when the strategy is created, so that the settings in the
subcloud group are used.
Alternatively, if several subclouds should be upgraded or updated individually,
they can be grouped together in a subcloud group that supports serial updates.
In this case, the :command:`max\_parallel\_subclouds`,
and:command:`subcloud\_apply\_type` are **not** specified when creating the
In this case, the :command:`max_parallel_subclouds`,
and:command:`subcloud_apply_type` are **not** specified when creating the
strategy, and the subcloud group settings for
:command:`max\_parallel\_subclouds` \(not applicable\), and the
:command:`subcloud\_apply\_type` \(serial\) associated with that subcloud group
:command:`max_parallel_subclouds` (not applicable), and the
:command:`subcloud_apply_type` (serial) associated with that subcloud group
are used.
For more information on creating a strategy for orchestration upgrades, updates
@ -44,7 +44,7 @@ or firmware updates, see:
Upgrade Orchestration Process Using the CLI
<distributed-upgrade-orchestration-process-using-the-cli>`.
- To create an update \(patch\) orchestration strategy use the
- To create an update (patch) orchestration strategy use the
:command:`dcmanager patch-strategy create` command.
.. xbooklink For more information see,

View File

@ -21,7 +21,7 @@ subcloud groups, is the order they are processed by orchestration.
#. All subclouds in the first, second, and third group, etc.
#. Subclouds from different groups will never be included in the same stage of
the strategy to ensure they are not upgraded, updated \(patched\) at the
the strategy to ensure they are not upgraded, updated (patched) at the
same time.

View File

@ -23,7 +23,7 @@ Controller using the rehoming playbook.
and the subcloud's controllers before rehoming the subcloud.
Use the following procedure to enable subcloud rehoming and to update the new
subcloud configuration \(networking parameters, passwords, etc.\) to be
subcloud configuration (networking parameters, passwords, etc.) to be
compatible with the new System Controller.
.. rubric:: |context|
@ -63,7 +63,7 @@ There are six phases for Rehoming a subcloud:
.. rubric:: |prereq|
- Ensure that the subcloud management subnet, oam_floating_address,
oam_node_0_address and oam_node_1_address \(if applicable\) does not overlap
oam_node_0_address and oam_node_1_address (if applicable) does not overlap
addresses already being used by the new System Controller or any of its
subclouds.
@ -130,7 +130,7 @@ There are six phases for Rehoming a subcloud:
You will need to specify the old and the new password.
#. For an |AIO-DX| subcloud, ensure that the active controller is
controller-0. Perform a host-swact of the active controller \(controller-1\)
controller-0. Perform a host-swact of the active controller (controller-1)
to make controller-0 active.
.. code-block:: none

View File

@ -9,7 +9,7 @@ Reinstall a Subcloud with Redfish Platform Management Service
=============================================================
For subclouds with servers that support Redfish Virtual Media Service
\(version 1.2 or higher\), you can use the Central cloud's CLI to reinstall
\(version 1.2 or higher), you can use the Central cloud's CLI to reinstall
the ISO and bootstrap subclouds from the Central cloud.
.. caution::

View File

@ -33,12 +33,12 @@ subclouds.
The Patch State indicates whether the patch is available,
partially-applied or applied. Applied indicates that the update has
been installed on all hosts of the cloud \(SystemController in this
case\).
been installed on all hosts of the cloud (SystemController in this
case).
#. Check the Update Sync Status of the subclouds.
Update \(or Patch\) Sync Status is part of the overall Sync status of a
Update (or Patch) Sync Status is part of the overall Sync status of a
subcloud. To review the synchronization status of subclouds, see
:ref:`Monitoring Subclouds Using Horizon
<monitoring-subclouds-using-horizon>`.

View File

@ -37,9 +37,9 @@ Distributed Cloud Using Horizon
The **Patch State** column indicates whether the Patch is available,
partially-applied or applied. **Applied** indicates that the update has
been installed on all hosts of the cloud \(SystemController in this case\).
been installed on all hosts of the cloud (SystemController in this case).
- To identify which subclouds are update-current \(**in-sync**\), use the
- To identify which subclouds are update-current (**in-sync**), use the
.. code-block:: none

View File

@ -13,10 +13,10 @@ Synchronizations can be delayed slightly, depending on network traffic
conditions and the amount of information to be synchronized.
|prod| synchronizes configuration for selected attributes of system-wide
configurations \(see :ref:`Table 1
<shared-configurations-shared-sys-configs>`\) and synchronizes configuration
for resources of the Keystone Identity Service \(see :ref:`Table 2
<shared-configurations-shared-keystone-configs>`\).
configurations (see :ref:`Table 1
<shared-configurations-shared-sys-configs>`) and synchronizes configuration
for resources of the Keystone Identity Service (see :ref:`Table 2
<shared-configurations-shared-keystone-configs>`).
.. _shared-configurations-shared-sys-configs:

View File

@ -20,7 +20,7 @@ and to collect alarms from the subcloud.
The subcloud is synchronized when it is first connected to the |prod-dc| and
set to managed. A backup audit and synchronization is run at regular intervals
\(every ten minutes\) for subclouds in the Managed state, synchronizing them to
\(every ten minutes) for subclouds in the Managed state, synchronizing them to
the System Controller. You can view the synchronization status for individual
subclouds on the Cloud Overview page from **Distributed Cloud Admin** \>
**Cloud Overview**.

View File

@ -6,11 +6,11 @@
Update Management for Distributed Cloud
=======================================
You can apply software updates \(also known as 'patches'\) to the Central Cloud
You can apply software updates (also known as 'patches') to the Central Cloud
and subclouds from the System Controller.
A central update repository on the Central Cloud is introduced for |prod-dc|.
This is used to store all updates \(patches\) so that unmanaged subclouds can
This is used to store all updates (patches) so that unmanaged subclouds can
be synchronized with any required updates when they are brought into a managed
state.

View File

@ -43,18 +43,18 @@ available:
parallel, or serially.
If this is not specified using the |CLI|, the values for
:command:`subcloud\_update\_type` defined for each subcloud group will be
:command:`subcloud_update_type` defined for each subcloud group will be
used by default.
**maximum parallel subclouds**
Sets the maximum number of subclouds that can be updated in parallel \(default 20\).
Sets the maximum number of subclouds that can be updated in parallel (default 20).
If this is not specified using the |CLI|, the values for
:command:`max\_parallel\_subclouds` defined for each subcloud group will be
:command:`max_parallel_subclouds` defined for each subcloud group will be
used by default.
**stop on failure**
true \(default\) or false — determines whether update orchestration failure
true (default) or false — determines whether update orchestration failure
for a subcloud prevents application to subsequent subclouds.
@ -224,7 +224,7 @@ individual subclouds.
.. note::
Since re-location is not possible on a single-node |prod| Simplex system,
you must change the configuration to set default\_instance\_action to
you must change the configuration to set default_instance_action to
stop-start.
.. _update-orchestration-of-central-clouds-regionone-and-subclouds-using-the-cli-ul-xfb-bfz-fdb:
@ -307,7 +307,7 @@ individual subclouds.
migrate or stop-start — determines whether hosted application VMs are
migrated or stopped and restarted when a worker host is upgraded.
**subcloud\_name**
**subcloud_name**
The name of the subcloud to use the custom strategy. If this omitted,
the default update strategy is updated.

View File

@ -33,7 +33,7 @@ Follow the steps below to manually upgrade the System Controller:
:start-after: license-begin
:end-before: license-end
#. Transfer iso and signature files to controller-0 \(active controller\) and import the load.
#. Transfer iso and signature files to controller-0 (active controller) and import the load.
.. code-block:: none
@ -158,7 +158,7 @@ Follow the steps below to manually upgrade the System Controller:
- State entered after :command:`system upgrade-start` completes.
- Release <nn.nn> system data \(for example, postgres databases\) has
- Release <nn.nn> system data (for example, postgres databases) has
been exported to be used in the upgrade.
As part of the upgrade, the upgrade process checks the health of the system
@ -366,10 +366,10 @@ Follow the steps below to manually upgrade the System Controller:
you can safely unlock the node.
After upgrading a storage node, but before unlocking, there are Ceph
synchronization alarms \(that appear to be making progress in
synching\), and there are infrastructure network interface alarms
\(since the infrastructure network interface configuration has not been
applied to the storage node yet, as it has not been unlocked\).
synchronization alarms (that appear to be making progress in
synching), and there are infrastructure network interface alarms
(since the infrastructure network interface configuration has not been
applied to the storage node yet, as it has not been unlocked).
Unlock the node as soon as the upgraded storage node comes online.
@ -447,8 +447,8 @@ Follow the steps below to manually upgrade the System Controller:
+--------------+--------------------------------------+
During the running of the :command:`upgrade-activate` command, new
configurations are applied to the controller. 250.001 \(**hostname
Configuration is out-of-date**\) alarms are raised and are cleared as the
configurations are applied to the controller. 250.001 (**hostname
Configuration is out-of-date**) alarms are raised and are cleared as the
configuration is applied. The upgrade state goes from **activating** to
**activation-complete** once this is done.

View File

@ -6,7 +6,7 @@
Upload and Applying Updates to SystemController Using Horizon
=============================================================
You can upload and apply updates \(patches\) to the SystemController in order
You can upload and apply updates (patches) to the SystemController in order
to update the central update repository, from the Horizon Web interface.
.. rubric:: |context|
@ -30,7 +30,7 @@ and Applying Updates to SystemController Using the CLI
#. On the **Patches** tab, click **Upload Patches**.
In the **Upload Patches** dialog box, click **Browse** to select updates
\(patches\) for upload.
(patches) for upload.
.. image:: figures/cah1525101473925.png

View File

@ -21,7 +21,7 @@ patches for the SystemController is provided below.
For standard |prod| updating procedures, see the |updates-doc|:
:ref:`software-updates-and-upgrades-software-updates` guide.
For SystemController of |prod-dc| \(and the central update repository\), you
For SystemController of |prod-dc| (and the central update repository), you
must include the additional |CLI| parameter ``--os-region-name`` with the value
SystemController when using |CLI| :command:`sw-patch` commands.
@ -67,7 +67,7 @@ SystemController when using |CLI| :command:`sw-patch` commands.
You may receive a warning about the update already being imported. This
is expected and occurs if the update was uploaded locally to the system
controller. The warning will only occur for patches that were applied
to controller-0 \(system controller\) before it was first unlocked.
to controller-0 (system controller) before it was first unlocked.
#. Confirm that the newly uploaded patches have a status of **available**.

View File

@ -530,7 +530,7 @@ traps will not be generated.
#. Modify your |SNMP| Helm chart values file (for example, ``user_conf.yaml``)
by adding the line "trap-server-port: [new port]" as shown in the example
below \("30162" is the new port in this example\).
below ("30162" is the new port in this example).
.. code-block:: none

View File

@ -27,7 +27,7 @@ system application.
About SNMP Support
------------------
Support for Simple Network Management Protocol \(SNMP\) is implemented as follows:
Support for Simple Network Management Protocol (SNMP) is implemented as follows:
.. _snmp-overview-ul-bjv-cjd-cp:
@ -56,7 +56,7 @@ For information on enabling SNMP support, see
.. _snmp-overview-section-N10099-N1001F-N10001:
-----------------------
SNMPv2-MIB \(RFC 3418\)
SNMPv2-MIB (RFC 3418)
-----------------------
Support for the basic standard MIB for SNMP entities is limited to the System
@ -86,17 +86,17 @@ Wind River Enterprise MIBs
|prod| supports the Wind River Enterprise Registration and Alarm MIBs.
**Enterprise Registration MIB, wrsEnterpriseReg.mib**
Defines the Wind River Systems \(WRS\) hierarchy underneath the
**iso\(1\).org\(3\).dod\(6\).internet\(1\).private\(4\).enterprise\(1\)**.
Defines the Wind River Systems (WRS) hierarchy underneath the
**iso\(1).org\(3).dod\(6).internet\(1).private\(4).enterprise\(1)**.
This hierarchy is administered as follows:
- **.wrs\(731\)**, the IANA-registered enterprise code for Wind River
- **.wrs\(731)**, the IANA-registered enterprise code for Wind River
Systems
- **.wrs\(731\).wrsCommon\(1\).wrs<Module\>\(1-...\)**,
- **.wrs\(731).wrsCommon\(1).wrs<Module\>\(1-...)**,
defined in wrsCommon<Module\>.mib.
- **.wrs\(731\).wrsProduct\(2-...\)**, defined in wrs<Product\>.mib.
- **.wrs\(731).wrsProduct\(2-...)**, defined in wrs<Product\>.mib.
**Alarm MIB, wrsAlarmMib.mib**
Defines the common TRAP and ALARM MIBs for |org| products.

View File

@ -37,8 +37,8 @@ unnecessary alarms.
<alarm-id>: **Alarm ID not found: <alarm-id\>**.
If the specified number of Alarm IDs is greater than 1, and at least 1 is
wrong, then the suppress command is not applied \(none of the specified
Alarm IDs are suppressed\).
wrong, then the suppress command is not applied (none of the specified
Alarm IDs are suppressed).
.. note::
Suppressing an Alarm will result in the system NOT notifying the

View File

@ -36,7 +36,7 @@ customer logs are mapped into the 'Message' trap.
MIBs. See :ref:`SNMP Overview <snmp-overview>` for details.
For Critical, Major, Minor, Warning, and Message traps, all variables in the
active alarm table are included as varbinds \(variable bindings\), where each
active alarm table are included as varbinds (variable bindings), where each
varbind is a pair of fields consisting of an object identifier and a value
for the object.

View File

@ -38,7 +38,7 @@ Collect Tool Caveats and Usage
- For |prod| Standard systems, use the following commands:
- For a small deployment \(less than two worker nodes\):
- For a small deployment (less than two worker nodes):
.. code-block:: none

View File

@ -32,7 +32,7 @@ If you need to reactivate a suppressed alarm, you can do so using the CLI.
``--uuid``
includes the alarm type UUIDs in the output.
Alarm type\(s\) with the specified <alarm-id\(s\)> will be unsuppressed.
Alarm type\(s) with the specified <alarm-id\(s)> will be unsuppressed.
You can unsuppress all currently suppressed alarms using the following command:

View File

@ -109,7 +109,7 @@ To review detailed information about a specific alarm instance, see
This option indicates that all active alarms should be displayed,
including suppressed alarms. Suppressed alarms are displayed with
their Alarm ID set to S<\(alarm-id\)>.
their Alarm ID set to S<\(alarm-id)>.
**--uuid**
The ``--uuid`` option on the :command:`fm alarm-list` command lists the

View File

@ -41,16 +41,16 @@ You can view detailed information to help troubleshoot an alarm.
| uuid | 4ab5698a-19cb-4c17-bd63-302173fef62c |
+------------------------+-------------------------------------------------+
The pair of attributes **\(alarm\_id, entity\_instance\_id\)** uniquely
The pair of attributes **\(alarm_id, entity_instance_id)** uniquely
identifies an active alarm:
**alarm\_id**
**alarm_id**
An ID identifying the particular alarm condition. Note that there are
some alarm conditions, such as *administratively locked*, that can be
raised by more than one entity-instance-id.
**entity\_instance\_id**
**entity_instance_id**
Type and instance information of the object raising the alarm. A
period-separated list of \(key, value\) pairs, representing the
period-separated list of (key, value) pairs, representing the
containment structure of the overall entity instance. This structure
is used for processing hierarchical clearing of alarms.

View File

@ -33,7 +33,7 @@ You can use CLI commands to work with historical alarms and logs in the event lo
Optional arguments:
``-q QUERY, --query QUERY``
\- key\[op\]data\_type::value; list. data\_type is optional, but if
\- key\[op\]data_type::value; list. data_type is optional, but if
supplied must be string, integer, float, or boolean.
``-l NUMBER, --limit NUMBER``
@ -45,7 +45,7 @@ You can use CLI commands to work with historical alarms and logs in the event lo
``--logs``
Show customer logs only.
``--include\_suppress``
``--include_suppress``
Show suppressed alarms as well as unsuppressed alarms.
``--uuid``

View File

@ -49,7 +49,7 @@ The following prerequisites are required before the integration:
:start-after: prereq-begin
:end-before: prereq-end
- The cloud is configured with a node that supports the Subordinate mode \(Secondary mode\).
- The cloud is configured with a node that supports the Subordinate mode (Secondary mode).
- The cloud is labeled with **ptp-registration=true**, and **ptp-notification=true**.

View File

@ -96,8 +96,8 @@ Config drive
------------
|prod-os| can be configured to use a special-purpose configuration
drive \(abbreviated config drive\) to store metadata \(including
injected files\). Metadata is written to the drive, which is attached
drive (abbreviated config drive) to store metadata (including
injected files). Metadata is written to the drive, which is attached
to the instance when it boots. The instance can retrieve information
normally available through the metadata service by reading from the
mounted drive.

View File

@ -25,7 +25,7 @@ hosting compute or |AIO|-controller node, do the following in the |VM|:
~(keystone_admin)$ modprobe kvm_ptp
#. Update the reference clock in the chrony config file \(/etc/chrony.conf\):
#. Update the reference clock in the chrony config file (/etc/chrony.conf):
.. code-block:: none

View File

@ -103,7 +103,7 @@ Where:
- OpenStack Horizon - dashboard
.. - Telemetry \(OPTIONAL\)
.. - Telemetry (OPTIONAL)
- Panko - Event storage

View File

@ -12,9 +12,9 @@ maintenance purposes.
On a controller node, the state transition only succeeds if there are no
services running in active mode on the host.
On a worker node \(or |AIO| Controller\), the state
On a worker node (or |AIO| Controller), the state
transition only succeeds if all currently running containers
\(hosted applications\) on the host can be re-located on alternative worker
\(hosted applications) on the host can be re-located on alternative worker
nodes or |AIO| Controller. Re-location of containers is
initiated automatically by |prod| as soon as the state transition is requested.
For containers, a live re-location of the container to another host is

View File

@ -14,8 +14,8 @@ maintenance purposes.
On a controller node, the state transition only succeeds if there are no
services running in active mode on the host.
On a worker node \(or |AIO|\), the state transition only succeeds if all
currently running containers \(hosted applications\) on the host can be
On a worker node (or |AIO|), the state transition only succeeds if all
currently running containers (hosted applications) on the host can be
re-located on alternative worker nodes or |AIO| Controller. Re-location of
containers is initiated automatically by |prod| as soon as the state transition
is requested. For containers, a live re-location of the container to

View File

@ -35,7 +35,7 @@ order. Observe the following tips:
Typically, this is the disk associated with /dev/sda and is as
defined in /proc/cmdline when the load is booted.
#. The |NIC| on the boot interface \(such as management or |PXE| network\).
#. The |NIC| on the boot interface (such as management or |PXE| network).
- Set the BIOS boot options to ensure a failsafe boot, if available. For
example, rotating through available boot interfaces, watchdog timer on

View File

@ -9,7 +9,7 @@ Swact Controllers Using Horizon
Swacting initiates a switch of the active/standby roles between two
controllers.
Swact is an abbreviated form of the term Switch Active \(host\). When
Swact is an abbreviated form of the term Switch Active (host). When
selected, this option forces the other controller to become the active one in
the HA cluster. This means that all active system services on this controller
move to standby operation, and that the corresponding services on the other

View File

@ -11,7 +11,7 @@ controllers.
.. rubric:: |context|
Swact is an abbreviated form of the term Switch Active \(host\). When
Swact is an abbreviated form of the term Switch Active (host). When
selected, this option forces the other controller to become the active one
in the HA cluster. This means that all active system services on this
controller move to standby operation, and that the corresponding services

View File

@ -23,7 +23,7 @@ successive reboot attempts is limited; for more information,
see :ref:`reboot-limits-for-host-unlock-d9a26854590a`.
.. note::
On a |prod| Simplex system, any containers \(hosted applications\) that
On a |prod| Simplex system, any containers (hosted applications) that
were stopped by the preceding **Lock Host** operation are started
automatically.

View File

@ -23,7 +23,7 @@ number of successive reboot attempts is limited; for more information,
see :ref:`reboot-limits-for-host-unlock-d9a26854590a`.
.. note::
On a |prod| Simplex system, any containers \(hosted applications\)
On a |prod| Simplex system, any containers (hosted applications)
that were stopped by the preceding :command:`host-lock` operation are
started automatically.

View File

@ -12,10 +12,10 @@ The hyper-threading status is controlled by the BIOS settings of the host.
Some applications may benefit from hyperthreading. For applications that
require deterministic performance, it is recommended to run with
hyperthreading disabled. If hyperthreading is enabled, the application
\(either running on bare metal or in a container\) must check the CPU
(either running on bare metal or in a container) must check the CPU
topology for the CPUs and affine tasks appropriately to HT siblings. For
example, "/proc/cpuinfo" and
"/sys/devices/system/cpu/cpuX/topology/thread\_siblings\*" can be used to
"/sys/devices/system/cpu/cpuX/topology/thread_siblings\*" can be used to
identify HT siblings of the same core.
@ -39,7 +39,7 @@ The hyper-threading status is controlled by the BIOS settings of the host.
.. note::
Changes to the host's BIOS must be made while it is locked and it
must not be subsequently unlocked until it comes back online
\(locked-disabled-online\) and the updated Hyperthreading settings
(locked-disabled-online) and the updated Hyperthreading settings
are available in the inventory.
#. Boot the host in BIOS mode.

View File

@ -62,7 +62,7 @@ them more CPU cores from the Horizon Web interface.
increased flexibility for high-performance configurations. For
example, you can dedicate certain |NUMA| nodes for platform
use such that other |NUMA| nodes that service IRQ requests are left
available for the containers \(hosted applications\) that require
available for the containers (hosted applications) that require
high-performance IRQ servicing.
.. note::

View File

@ -46,17 +46,17 @@ see :ref:`The Life Cycle of a Host <the-life-cycle-of-a-host-93640aa2b707>`.
The following service parameters control the boot timeout interval.
**worker\_boot\_timeout**
**worker_boot_timeout**
The time in seconds to allow for a worker or storage host to boot
\(7201800 seconds\). The default value is 720 seconds \(12 minutes\).
(7201800 seconds). The default value is 720 seconds (12 minutes).
.. note::
This parameter also applies to storage nodes.
**controller\_boot\_timeout**
**controller_boot_timeout**
The time in seconds to allow for a controller host to boot
\(1200-1800 seconds\). The default value is 1200 seconds
\(20 minutes\).
(1200-1800 seconds). The default value is 1200 seconds
(20 minutes).
For example, to change the boot timeout for the worker and storage
hosts to 840 seconds:

View File

@ -10,13 +10,13 @@ You can adjust the heartbeat interval, as well as the thresholds for missed
heartbeat challenges that cause a host to be moved to the **Degraded** or
**Failed** state.
The settings apply to all hosts \(controller, worker, and storage\). For more
The settings apply to all hosts (controller, worker, and storage). For more
information about host states,
see :ref:`The Life Cycle of a Host <the-life-cycle-of-a-host-93640aa2b707>`.
.. note::
The heartbeat\_degrade threshold must not exceed the
heartbeat\_failure\_threshold.
The heartbeat_degrade threshold must not exceed the
heartbeat_failure_threshold.
.. rubric:: |proc|
@ -51,19 +51,19 @@ see :ref:`The Life Cycle of a Host <the-life-cycle-of-a-host-93640aa2b707>`.
response thresholds for moving a host to the **Degraded** or **Failed**
state.
**heartbeat\_period**
**heartbeat_period**
The time in milliseconds between heartbeat challenges from the
controller to the other hosts \(1001000 ms\). The default is
controller to the other hosts (1001000 ms). The default is
100 ms.
**heartbeat\_degrade\_threshold**
**heartbeat_degrade_threshold**
The number of consecutive missing responses to heartbeat challenges
before a host is moved into the **Degraded** state \(4100\). The
before a host is moved into the **Degraded** state (4100). The
default is six consecutive missing responses.
**heartbeat\_failure\_threshold**
**heartbeat_failure_threshold**
The number of consecutive missing responses to heartbeat challenges
before a host is moved into the **Failed** state \(10100\). The
before a host is moved into the **Failed** state (10100). The
default is 10 consecutive missing responses.
For example, to change the heartbeat failure threshold for all hosts to

View File

@ -6,7 +6,7 @@
Configure Heartbeat Failure Action
==================================
You can configure **heartbeat\_failure\_action** while performing network
You can configure **heartbeat_failure_action** while performing network
related maintenance activities that may interrupt inter-host communications.
You can configure service parameters to change the heartbeat failure behavior
@ -50,7 +50,7 @@ immediately in the event of a persistent loss of maintenance heartbeat.
~(keystone_admin)$ system service-parameter-modify <platform maintenance heartbeat_failure_action>=ignore Action must be one of 'fail', 'degrade', 'alarm' or 'none'
The following service parameters control the
**heartbeat\_failure\_action** and accepts one of the four possible
**heartbeat_failure_action** and accepts one of the four possible
actions.
**fail**
@ -121,11 +121,11 @@ The heartbeat alarms, such as Management Network can be viewed. For example:
.. note::
In the event of a single host heartbeat failure, maintenance will attempt
to reboot, and if unreachable, will also attempt to reset the host in order
to expedite failed host recovery (if |LAG| Network is provisioned\).
to expedite failed host recovery (if |LAG| Network is provisioned).
.. warning::
To maintain a system with High Fault Detection and Availability the
**heartbeat\_failure\_action** should always be reverted back to **fail**
**heartbeat_failure_action** should always be reverted back to **fail**
once network maintenance activities are completed. This action applies to
all hosts and if a heartbeat failure occurs while any action other than
**fail** is selected, maintenance will not take action to recover the
@ -143,5 +143,5 @@ The heartbeat alarms, such as Management Network can be viewed. For example:
.. rubric:: |postreq|
Always revert the **heartbeat\_failure\_action** to **fail** once network
Always revert the **heartbeat_failure_action** to **fail** once network
maintenance activities are complete.

View File

@ -11,19 +11,19 @@ maintenance heartbeat failures of more than one host, and gracefully
recovers the hosts once the heartbeat is re-established.
You can configure multi-node failure avoidance for recovery of failing hosts,
**mnfa\_threshold** \(default is 2, range is specified from 2 to 100\), and
**mnfa_threshold** (default is 2, range is specified from 2 to 100), and
the number of seconds the heartbeat can fail in this group of hosts,
**mnfa\_timeout** \(default is no-timeout, value of 0, or from 100 to 86400
secs=1 day\), before the hosts are declared failed, or are required to be
**mnfa_timeout** (default is no-timeout, value of 0, or from 100 to 86400
secs=1 day), before the hosts are declared failed, or are required to be
forced to reboot/reset. If the value is set outside the range, a warning is
displayed.
Multi-Node Failure Avoidance is based on four or more back to back heartbeat
pulse misses for a **mnfa\_threshold** or higher number of hosts within a
pulse misses for a **mnfa_threshold** or higher number of hosts within a
full heartbeat loss window. For example, given the default heartbeat period
of 100 msec and the **heartbeat\_failure\_threshold** of 10; if maintenance
sees **mnfa\_threshold** or more hosts missing four or more back to back
heartbeat responses within one second \( 100 msec times 10 \), then
of 100 msec and the **heartbeat_failure_threshold** of 10; if maintenance
sees **mnfa_threshold** or more hosts missing four or more back to back
heartbeat responses within one second ( 100 msec times 10 ), then
Multi-Node Failure Avoidance is activated for those hosts. Any additional
hosts failing heartbeat while |MNFA| is active are added to the |MNFA| pool.
@ -31,23 +31,23 @@ In Horizon, |MNFA| displays heartbeat failing hosts in the
**unlocked-enabled-degraded** state, and displays a status of “Graceful
Recovery Wait” while maintenance waits for heartbeat to that host to recover.
This degraded state and host status is true only for the **fail** and
**degrade** **heartbeat\_failure\_action** selections. For information on
**degrade** **heartbeat_failure_action** selections. For information on
viewing heartbeat-failing hosts from Horizon, see :ref:`Hosts Tab <hosts-tab>`.
Hosts whose heartbeat recovers, after ten back to back heartbeat responses,
are removed from the |MNFA| pool with state and status returned to what it was
prior to the event. Once the |MNFA| pool size drops below **mnfa\_threshold**,
then the remaining hosts have 6 seconds \(100 msec times 10 plus 5 second grace
period\) to recover before the selected **heartbeat\_failure\_action** is taken
against the hosts. With the **mnfa\_threshold** of two that would only be one
host \(or for 3 that could be 2\). If late recovering hosts recover, and if
prior to the event. Once the |MNFA| pool size drops below **mnfa_threshold**,
then the remaining hosts have 6 seconds (100 msec times 10 plus 5 second grace
period) to recover before the selected **heartbeat_failure_action** is taken
against the hosts. With the **mnfa_threshold** of two that would only be one
host (or for 3 that could be 2). If late recovering hosts recover, and if
their uptime shows that they had rebooted, then they are tested and brought
back into service like the others. Otherwise they are fully re-enabled
through reboot/reset.
In |MNFA| recovery, where the **heartbeat\_failure\_action** is **fail** and
In |MNFA| recovery, where the **heartbeat_failure_action** is **fail** and
the hosts do not reboot during the loss of heartbeat. It is possible that
maintenance will force a reboot of the **mnfa\_threshold-1** late recovering
maintenance will force a reboot of the **mnfa_threshold-1** late recovering
hosts upon eventual recovery, as at that point they are treated as individual
heartbeat failures.
@ -56,7 +56,7 @@ heartbeat failures.
.. _configuring-multi-node-failure-avoidance-steps-m4h-j3h-gfb:
#. Use the :command:`system service-parameter-modify` command to specify the
new **mnfa\_threshold** and **mnfa\_timeout** setting. Changing this to
new **mnfa_threshold** and **mnfa_timeout** setting. Changing this to
an invalid value, results in a semantic check error similar to the
following:
@ -65,8 +65,8 @@ heartbeat failures.
~(keystone_admin)$ system service-parameter-modify platform maintenance mnfa_threshold=<1>
Parameter 'mnfa_threshold' must be between 2 and 100
The **mnfa\_timeout** accepts a value of 0 indicating no-timeout or
from 100 to 86400 secs \(1 day\). For example:
The **mnfa_timeout** accepts a value of 0 indicating no-timeout or
from 100 to 86400 secs (1 day). For example:
.. code-block:: none
@ -88,7 +88,7 @@ heartbeat failures.
.. note::
Multi-Node Failure Avoidance is never activated as it does not apply
if the **heartbeat\_failure\_action** is set to **alarm** or **none**.
if the **heartbeat_failure_action** is set to **alarm** or **none**.
For more information, see :ref:`Configure Heartbeat Failure Action
<configuring-heartbeat-failure-action>`.

View File

@ -50,8 +50,8 @@ enables the ACC100/ACC200 device.
~(keystone_admin)$ system host-label-assign controller-0 kube-topology-mgr-policy=restricted
#. Modify the CPU core assignments for controller-0 to have 12
application-isolated physical cores \(24 virtual cores if hyper-threading
is supported and enabled on the processor\) on processor 0. Your specific
application-isolated physical cores (24 virtual cores if hyper-threading
is supported and enabled on the processor) on processor 0. Your specific
application(s) may need more or less cores.

View File

@ -10,7 +10,7 @@ To initiate device image updates on a host, use the
:command:`host-device-image-update` command.
.. note::
Modifying or deleting device labels for devices on the host \(if any\)
Modifying or deleting device labels for devices on the host (if any)
is not allowed while the update is in progress.
The command syntax is:

View File

@ -33,16 +33,16 @@ displayed in the list of host devices by running the following command:
+-------+----------+--------+--------+--------+------------+-------------+----------------------+-----------+---------|
To enable the |FEC| device for |SRIOV| interfaces, it must be modified in order
to set the number of virtual functions \(VF\), and the appropriate userspace
drivers for the physical function \(PF\), and VFs.
to set the number of virtual functions (VF), and the appropriate userspace
drivers for the physical function (PF), and VFs.
The following PF and VF drivers are supported:
.. _n3000-fpga-forward-error-correction-ul-klj-2zh-bmb:
- PF driver: igb\_uio
- PF driver: igb_uio
- VF driver: igb\_uio, vfio
- VF driver: igb_uio, vfio
For example, run the following commands:

View File

@ -23,7 +23,7 @@ The following procedure shows an example of launching a container image with
$ source /etc/platform/openrc ~(keystone_admin)$
#. Create a pod.yml file that requests 16 ACC100/ACC200 VFs
\(i.e. intel.com/intel_acc100_fec: '16'\)
(i.e. intel.com/intel_acc100_fec: '16')
.. code-block:: none

Some files were not shown because too many files have changed in this diff Show More