Fixed \_ as the output was not rendering correctly (pick r5 updates only)

Fixed Patchset 4 comments
Fixed Patchset 3 comments and added additional updates
Signed-off-by: Juanita-Balaraj <juanita.balaraj@windriver.com>
Change-Id: I7482afc3a90bbdc94b6ecd8b6ac39d831b8a45db
Signed-off-by: Juanita-Balaraj <juanita.balaraj@windriver.com>
This commit is contained in:
Juanita-Balaraj 2021-10-26 17:31:52 -04:00
parent 11bebe59c3
commit 265d96bed1
21 changed files with 113 additions and 113 deletions

View File

@ -51,7 +51,7 @@ commands to manage containerized applications provided as part of |prod|.
where:
**<app\_name>**
**<app_name>**
The name of the application to show details.
For example:
@ -83,7 +83,7 @@ commands to manage containerized applications provided as part of |prod|.
where the following are optional arguments:
**<app\_name>**
**<app_name>**
Assigns a custom name for application. You can use this name to
interact with the application in the future.
@ -92,7 +92,7 @@ commands to manage containerized applications provided as part of |prod|.
and the following is a positional argument:
**<tar\_file>**
**<tar_file>**
The path to the tar file containing the application to be uploaded.
For example:
@ -126,7 +126,7 @@ commands to manage containerized applications provided as part of |prod|.
where the following is a positional argument:
**<app\_name>**
**<app_name>**
The name of the application.
and the following is an optional argument:
@ -175,7 +175,7 @@ commands to manage containerized applications provided as part of |prod|.
+---------------------+--------------------------------+---------------+
- To show the overrides for a particular chart, use the following command.
System overrides are displayed in the **system\_overrides** section of
System overrides are displayed in the **system_overrides** section of
the **Property** column.
.. code-block:: none
@ -185,10 +185,10 @@ commands to manage containerized applications provided as part of |prod|.
where the following are positional arguments:
**<app\_name>**
**<app_name>**
The name of the application.
**< chart\_name>**
**< chart_name>**
The name of the chart.
**<namespace>**
@ -212,10 +212,10 @@ commands to manage containerized applications provided as part of |prod|.
where the following are positional arguments:
**<app\_name>**
**<app_name>**
The name of the application.
**<chart\_name>**
**<chart_name>**
The name of the chart.
**<namespace>**
@ -257,7 +257,7 @@ commands to manage containerized applications provided as part of |prod|.
| | DEBUG: true |
+----------------+-------------------+
The user overrides are shown in the **user\_overrides** section of the
The user overrides are shown in the **user_overrides** section of the
**Property** column.
.. note::
@ -280,10 +280,10 @@ commands to manage containerized applications provided as part of |prod|.
and the following are positional arguments:
**<app\_name>**
**<app_name>**
The name of the application.
**<chart\_name>**
**<chart_name>**
The name of the chart.
**<namespace>**
@ -302,10 +302,10 @@ commands to manage containerized applications provided as part of |prod|.
where the following are positional arguments:
**<app\_name>**
**<app_name>**
The name of the application.
**<chart\_name>**
**<chart_name>**
The name of the chart.
**<namespace>**
@ -334,7 +334,7 @@ commands to manage containerized applications provided as part of |prod|.
and the following is a positional argument:
**<app\_name>**
**<app_name>**
The name of the application to apply.
For example:
@ -366,7 +366,7 @@ commands to manage containerized applications provided as part of |prod|.
where:
**<app\_name>**
**<app_name>**
The name of the application to abort.
For example:
@ -389,7 +389,7 @@ commands to manage containerized applications provided as part of |prod|.
where the following are optional arguments:
**<app\_name>**
**<app_name>**
The name of the application to update.
You can look up the name of an application using the
@ -417,7 +417,7 @@ commands to manage containerized applications provided as part of |prod|.
and the following is a positional argument which must come last:
**<tar\_file>**
**<tar_file>**
The tar file containing the application manifest, Helm charts and
configuration file.
@ -431,7 +431,7 @@ commands to manage containerized applications provided as part of |prod|.
where:
**<app\_name>**
**<app_name>**
The name of the application to remove.
For example:
@ -466,7 +466,7 @@ commands to manage containerized applications provided as part of |prod|.
where:
**<app\_name>**
**<app_name>**
The name of the application to delete.
You must run :command:`application-remove` before deleting an application.

View File

@ -52,7 +52,7 @@ example, the user **testuser** is correct in the following URL, while
cloud/systemController, access to the local registry can only be done using
registry.local:9001. registry.central:9001 will be inaccessible. Installing
a |CA|-signed certificate for the registry and the certificate of the |CA| as
an 'ssl\_ca' certificate will remove this restriction.
an 'ssl_ca' certificate will remove this restriction.
For more information about Docker commands, see
`https://docs.docker.com/engine/reference/commandline/docker/ <https://docs.docker.com/engine/reference/commandline/docker/>`__.

View File

@ -128,7 +128,7 @@ conditions are in place:
#. Install network connectivity required for the subcloud.
#. Ensure that the backup file are available on the controller. Run both
Ansible Restore playbooks, restore\_platform.yml and restore\_user\_images.yml.
Ansible Restore playbooks, restore_platform.yml and restore_user_images.yml.
For more information on restoring the back up file, see :ref:`Run Restore
Playbook Locally on the Controller
<running-restore-playbook-locally-on-the-controller>`, and :ref:`Run
@ -139,7 +139,7 @@ conditions are in place:
The backup files contain the system data and updates.
#. If the backup file contains patches, Ansible Restore playbook
restore\_platform.yml will apply the patches and prompt you to reboot the
restore_platform.yml will apply the patches and prompt you to reboot the
system, you will need to re-run Ansible Restore playbook.
The current software version on the controller is compared against the
@ -162,7 +162,7 @@ conditions are in place:
Rerun the Ansible Playbook if there were patches applied and you were
prompted to reboot the system.
#. Restore the local registry using the file restore\_user\_images.yml.
#. Restore the local registry using the file restore_user_images.yml.
This must be done before unlocking controller-0.
@ -176,7 +176,7 @@ conditions are in place:
becomes operational.
#. If the system is a Distributed Cloud system controller, restore the **dc-vault**
using the restore\_dc\_vault.yml playbook. Perform this step after unlocking
using the restore_dc_vault.yml playbook. Perform this step after unlocking
controller-0:
.. code-block:: none
@ -261,7 +261,7 @@ conditions are in place:
| 6 | compute-1 | worker | locked |disabled |offline |
+----+-------------+------------+---------------+-----------+------------+
#. Restore storage configuration. If :command:`wipe\_ceph\_osds` is set to
#. Restore storage configuration. If :command:`wipe_ceph_osds` is set to
**True**, follow the same procedure used to restore **controller-1**,
beginning with host **storage-0** and proceeding in sequence.
@ -275,12 +275,12 @@ conditions are in place:
the restore procedure without interruption.
Standard with Controller Storage install or reinstall depends on the
:command:`wipe\_ceph\_osds` configuration:
:command:`wipe_ceph_osds` configuration:
#. If :command:`wipe\_ceph\_osds` is set to **true**, reinstall the
#. If :command:`wipe_ceph_osds` is set to **true**, reinstall the
storage hosts.
#. If :command:`wipe\_ceph\_osds` is set to **false** \(default
#. If :command:`wipe_ceph_osds` is set to **false** \(default
option\), do not reinstall the storage hosts.
.. caution::
@ -315,9 +315,9 @@ conditions are in place:
.. caution::
Do not proceed until the Ceph cluster is healthy and the message
HEALTH\_OK appears.
HEALTH_OK appears.
If the message HEALTH\_WARN appears, wait a few minutes and then try
If the message HEALTH_WARN appears, wait a few minutes and then try
again. If the warning condition persists, consult the public
documentation for troubleshooting Ceph monitors \(for example,
`http://docs.ceph.com/docs/master/rados/troubleshooting/troubleshootin

View File

@ -15,7 +15,7 @@ Use the following command to run the Ansible Backup playbook and back up the
~(keystone_admin)]$ ansible-playbook /usr/share/ansible/stx-ansible/playbooks/backup.yml -e "ansible_become_pass=<sysadmin password> admin_password=<sysadmin password>" -e "backup_user_local_registry=true"
The <admin\_password> and <ansible\_become\_pass\> need to be set correctly
The <admin_password> and <ansible_become_pass> need to be set correctly
using the ``-e`` option on the command line, or an override file, or in the
Ansible secret file.
@ -23,44 +23,44 @@ The output files will be named:
.. _running-ansible-backup-playbook-locally-on-the-controller-ul-wj1-vxh-pmb:
- inventory\_hostname\_platform\_backup\_timestamp.tgz
- inventory_hostname_platform_backup_timestamp.tgz
- inventory\_hostname\_openstack\_backup\_timestamp.tgz
- inventory_hostname_openstack_backup_timestamp.tgz
- inventory\_hostname\_docker\_local\_registry\_backup\_timestamp.tgz
- inventory_hostname_docker_local_registry_backup_timestamp.tgz
- inventory\_hostname\_dc\_vault\_backup\_timestamp.tgz
- inventory_hostname_dc_vault_backup_timestamp.tgz
The variables prefix can be overridden using the ``-e`` option on the command
line or by using an override file.
.. _running-ansible-backup-playbook-locally-on-the-controller-ul-rdp-gyh-pmb:
- platform\_backup\_filename\_prefix
- platform_backup_filename_prefix
- openstack\_backup\_filename\_prefix
- openstack_backup_filename_prefix
- docker\_local\_registry\_backup\_filename\_prefix
- docker_local_registry_backup_filename_prefix
- dc\_vault\_backup\_filename\_prefix
- dc_vault_backup_filename_prefix
The generated backup tar files will be displayed in the following format,
for example:
.. _running-ansible-backup-playbook-locally-on-the-controller-ul-p3b-f13-pmb:
- localhost\_docker\_local\_registry\_backup\_2020\_07\_15\_21\_24\_22.tgz
- localhost_docker_local_registry_backup_2020_07_15_21_24_22.tgz
- localhost\_platform\_backup\_2020\_07\_15\_21\_24\_22.tgz
- localhost_platform_backup_2020_07_15_21_24_22.tgz
- localhost\_openstack\_backup\_2020\_07\_15\_21\_24\_22.tgz
- localhost_openstack_backup_2020_07_15_21_24_22.tgz
- localhost\_dc\_vault\_backup\_2020\_07\_15\_21\_24\_22.tgz
- localhost_dc_vault_backup_2020_07_15_21_24_22.tgz
These files are located by default in the /opt/backups directory on
controller-0, and contains the complete system backup.
If the default location needs to be modified, the variable backup\_dir can
If the default location needs to be modified, the variable backup_dir can
be overridden using the ``-e`` option on the command line or by using an
override file.

View File

@ -73,24 +73,24 @@ In this method you can run Ansible Restore playbook and point to controller-0.
expects both the **initial_backup_dir** and **backup_filename**
to be specified.
- The backup\_filename is the platform backup tar file. It must be
- The backup_filename is the platform backup tar file. It must be
provided using the ``-e`` option on the command line, for example:
.. code-block:: none
-e backup\_filename= localhost_platform_backup_2019_07_15_14_46_37.tgz
- The initial\_backup\_dir is the location on the Ansible control
- The initial_backup_dir is the location on the Ansible control
machine where the platform backup tar file is placed to restore the
platform. It must be provided using ``-e`` option on the command line.
- The :command:`admin\_password`, :command:`ansible\_become\_pass`,
and :command:`ansible\_ssh\_pass` need to be set correctly using
- The :command:`admin_password`, :command:`ansible_become_pass`,
and :command:`ansible_ssh_pass` need to be set correctly using
the ``-e`` option on the command line or in the Ansible secret file.
:command:`ansible\_ssh\_pass` is the password to the sysadmin user
:command:`ansible_ssh_pass` is the password to the sysadmin user
on controller-0.
- The :command:`ansible\_remote\_tmp` should be set to a new
- The :command:`ansible_remote_tmp` should be set to a new
directory \(not required to create it ahead of time\) under
/home/sysadmin on controller-0 using the ``-e`` option on the command
line.
@ -106,7 +106,7 @@ In this method you can run Ansible Restore playbook and point to controller-0.
the patches and prompt you to reboot the system. Then you will need to
re-run Ansible Restore playbook.
#. After running the restore\_platform.yml playbook, you can restore the local
#. After running the restore_platform.yml playbook, you can restore the local
registry images.
.. note::
@ -119,33 +119,33 @@ In this method you can run Ansible Restore playbook and point to controller-0.
where optional-extra-vars can be:
- The backup\_filename is the local registry backup tar file. It
- The backup_filename is the local registry backup tar file. It
must be provided using the ``-e`` option on the command line, for
example:
.. code-block:: none
-e backup\_filename= localhost_docker_local_registry_backup_2020_07_15_21_24_22.tgz
-e backup_filename= localhost_docker_local_registry_backup_2020_07_15_21_24_22.tgz
- The initial\_backup\_dir is the location on the Ansible control
- The initial_backup_dir is the location on the Ansible control
machine where the platform backup tar file is located. It must be
provided using ``-e`` option on the command line.
- The :command:`ansible\_become\_pass`, and
:command:`ansible\_ssh\_pass` need to be set correctly using the
- The :command:`ansible_become_pass`, and
:command:`ansible_ssh_pass` need to be set correctly using the
``-e`` option on the command line or in the Ansible secret file.
:command:`ansible\_ssh\_pass` is the password to the sysadmin user
:command:`ansible_ssh_pass` is the password to the sysadmin user
on controller-0.
- The backup\_dir should be set to a directory on controller-0.
- The backup_dir should be set to a directory on controller-0.
The directory must have sufficient space for local registry backup
to be copied. The backup\_dir is set using the ``-e`` option on the
to be copied. The backup_dir is set using the ``-e`` option on the
command line.
- The :command:`ansible\_remote\_tmp` should be set to a new
- The :command:`ansible_remote_tmp` should be set to a new
directory on controller-0. Ansible will use this directory to copy
files, and the directory must have sufficient space for local
registry backup to be copied. The :command:`ansible\_remote\_tmp`
registry backup to be copied. The :command:`ansible_remote_tmp`
is set using the ``-e`` option on the command line.
For example, run the local registry restore playbook, where

View File

@ -33,13 +33,13 @@ your server is isolated from the public Internet.
username: <your_my-registry.io_username>
password: <your_my-registry.io_password>
Where ``<your\_my-registry.io\_username>`` and
``<your\_my-registry.io\_password>`` are your login credentials for the
Where ``<your_my-registry.io_username>`` and
``<your_my-registry.io_password>`` are your login credentials for the
``<my-registry.io>`` private Docker registry.
.. note::
``<my-registry.io>`` must be a DNS name resolvable by the dns servers
configured in the ``dns\_servers:`` structure of the ansible bootstrap
configured in the ``dns_servers:`` structure of the ansible bootstrap
override file /home/sysadmin/localhost.yml.
#. For any additional local registry images required, use the full image name

View File

@ -33,13 +33,13 @@ your server is isolated from the public Internet.
username: <your_my-registry.io_username>
password: <your_my-registry.io_password>
Where ``<your\_my-registry.io\_username>`` and
``<your\_my-registry.io\_password>`` are your login credentials for the
Where ``<your_my-registry.io_username>`` and
``<your_my-registry.io_password>`` are your login credentials for the
``<my-registry.io>`` private Docker registry.
.. note::
``<my-registry.io>`` must be a DNS name resolvable by the dns servers
configured in the ``dns\_servers:`` structure of the ansible bootstrap
configured in the ``dns_servers:`` structure of the ansible bootstrap
override file /home/sysadmin/localhost.yml.
#. For any additional local registry images required, use the full image name

View File

@ -138,13 +138,13 @@ controller for access by subclouds. For example:
**--subcloud-apply-type**
Determines whether the subclouds are upgraded in parallel, or serially. If
this is not specified using the CLI, the values for subcloud\_update\_type
this is not specified using the CLI, the values for subcloud_update_type
defined for each subcloud group will be used by default.
**--max-parallel-subclouds**
Sets the maximum number of subclouds that can be upgraded in parallel
\(default 20\). If this is not specified using the CLI, the values for
max\_parallel\_subclouds defined for each subcloud group will be used by
max_parallel_subclouds defined for each subcloud group will be used by
default.
**--stop-on-failure**

View File

@ -35,8 +35,8 @@ You must be in **SystemController** mode. To change the mode, see
tab.
.. image:: figures/vhy1525122582274.png
#. On the Cloud Patching Orchestration tab, click **Create Strategy**.
@ -49,7 +49,7 @@ You must be in **SystemController** mode. To change the mode, see
parallel or serially.
If this is not specified using the |CLI|, the values for
:command:`subcloud\_update\_type` defined for each subcloud group will
:command:`subcloud_update_type` defined for each subcloud group will
be used by default.
**max-parallel-subclouds**
@ -57,7 +57,7 @@ You must be in **SystemController** mode. To change the mode, see
\(default 20\).
If this is not specified using the |CLI|, the values for
:command:`max\_parallel\_subclouds` defined for each subcloud group
:command:`max_parallel_subclouds` defined for each subcloud group
will be used by default.
**stop-on-failure**
@ -82,7 +82,7 @@ You must be in **SystemController** mode. To change the mode, see
To change the update strategy settings, you must delete the update
strategy and create a new one.
.. seealso::
.. seealso::
:ref:`Customizing the Update Configuration for Distributed Cloud Update
Orchestration <customizing-the-update-configuration-for-distributed-cloud-update-orchestration>`

View File

@ -54,7 +54,7 @@ device image updates, including |FPGA| updates.
.. code-block:: none
~(keystone_admin)]$ system host-device-list <hostname_or_id>
~(keystone_admin)]$ system host-device-list <hostname_or_id>
To list devices from the central cloud, run:
@ -143,7 +143,7 @@ device image updates, including |FPGA| updates.
parallel, or serially.
If this is not specified using the |CLI|, the values for
:command:`subcloud\_update\_type` defined for each subcloud group will
:command:`subcloud_update_type` defined for each subcloud group will
be used by default.
**max-parallel-subclouds**
@ -151,7 +151,7 @@ device image updates, including |FPGA| updates.
\(default 20\).
If this is not specified using the |CLI|, the values for
:command:`max\_parallel\_subclouds` defined for each subcloud group
:command:`max_parallel_subclouds` defined for each subcloud group
will be used by default.
**stop-on-failure**
@ -198,7 +198,7 @@ device image updates, including |FPGA| updates.
| created_at | 2020-08-11T18:13:40.576659 |
| updated_at | 2020-08-11T18:13:56.525459 |
+------------------------+----------------------------+
#. Monitor progress as the strategy is applied.
@ -217,7 +217,7 @@ device image updates, including |FPGA| updates.
+-----------+-------+----------+------------------------------+----------------------------+----------------------------+
| subcloud3 | 2 | applying | apply phase is 18% complete | 2020-08-13 14:12:13.457588 | None |
+-----------+-------+----------+------------------------------+----------------------------+----------------------------+
- To monitor the step currently being performed on a specific subcloud, do the following:

View File

@ -12,7 +12,7 @@ connected to the Central Cloud over L3 networks.
The Central Cloud has two regions: RegionOne, used to manage the nodes in the
Central Cloud, and System Controller, used to manage the subclouds in the
|prod-dc| system. You can select RegionOne or System Controller regions from the
Horizon Web interface or by setting the <OS\_REGION\_NAME> environment variable
Horizon Web interface or by setting the <OS_REGION_NAME> environment variable
if using the CLI.
**Central Cloud**

View File

@ -42,7 +42,7 @@ following conditions:
require remote install using Redfish.
- Redfish |BMC| is required for orchestrated subcloud upgrades. The install
values, and :command:`bmc\_password` for each |AIO-SX| subcloud controller
values, and :command:`bmc_password` for each |AIO-SX| subcloud controller
must be provided using the following |CLI| command on the System Controller.
.. note::
@ -84,7 +84,7 @@ following conditions:
each subcloud, use the following command to remove the previous upgrade
data:
:command:`sudo rm /opt/platform-backup/upgrade\_data\*`
:command:`sudo rm /opt/platform-backup/upgrade_data\*`
.. rubric:: |proc|
@ -93,7 +93,7 @@ following conditions:
#. Review the upgrade status for the subclouds.
After the System Controller upgrade is completed, wait for 10 minutes for
the **load\_sync\_status** of all subclouds to be updated.
the **load_sync_status** of all subclouds to be updated.
To identify which subclouds are upgrade-current \(in-sync\), use the
:command:`subcloud list` command. For example:
@ -169,7 +169,7 @@ following conditions:
upgraded in parallel, or serially.
If this is not specified using the CLI, the values for
:command:`subcloud\_update\_type` defined for each subcloud group will
:command:`subcloud_update_type` defined for each subcloud group will
be used by default.
**max-parallel-subclouds**
@ -177,7 +177,7 @@ following conditions:
\(default 20\).
If this is not specified using the CLI, the values for
:command:`max\_parallel\_subclouds` defined for each subcloud group
:command:`max_parallel_subclouds` defined for each subcloud group
will be used by default.
**stop-on-failure**

View File

@ -31,8 +31,8 @@ subcloud, the subcloud installation has these phases:
After a successful remote installation of a subcloud in a Distributed Cloud
system, a subsequent remote reinstallation fails because of an existing ssh
key entry in the /root/.ssh/known\_hosts on the System Controller. In this
case, delete the host key entry, if present, from /root/.ssh/known\_hosts
key entry in the /root/.ssh/known_hosts on the System Controller. In this
case, delete the host key entry, if present, from /root/.ssh/known_hosts
on the System Controller before doing reinstallations.
.. rubric:: |prereq|
@ -203,7 +203,7 @@ subcloud, the subcloud installation has these phases:
password: <sysinv_password>
type: docker
Where <sysinv\_password> can be found by running the following command as
Where <sysinv_password> can be found by running the following command as
'sysadmin' on the Central Cloud:
.. code-block:: none
@ -230,7 +230,7 @@ subcloud, the subcloud installation has these phases:
If you prefer to install container images from the default WRS |AWS| ECR
external registries, make the following substitutions for the
**docker\_registries** sections of the file.
**docker_registries** sections of the file.
.. code-block:: none
@ -332,9 +332,9 @@ subcloud, the subcloud installation has these phases:
bootstrapping and deployment by monitoring the following log files on the
active controller in the Central Cloud.
/var/log/dcmanager/<subcloud\_name>\_install\_<date\_stamp>.log.
/var/log/dcmanager/<subcloud_name>_install_<date_stamp>.log.
/var/log/dcmanager/<subcloud\_name>\_bootstrap\_<date\_stamp>.log.
/var/log/dcmanager/<subcloud_name>_bootstrap_<date_stamp>.log.
For example:

View File

@ -29,7 +29,7 @@ configuration file.
To send user data when calling nova boot, use the ``--user-data
/path/to/filename`` option, or use the Heat service and set the
``user\_data`` property and ``user\_data\_format`` to RAW.
``user_data`` property and ``user_data_format`` to RAW.
On initialization, the |VM| queries the metadata service through either
the EC2 compatibility API. For example:

View File

@ -82,7 +82,7 @@ They can only be modified to use different physical ports when required.
``ifname``
The name of the interface.
``ip\_address``
``ip_address``
An IPv4 or IPv6 address.
``prefix``

View File

@ -49,7 +49,7 @@ Remarks
:command:`system modify --description="A TEST"` is logged to sysinv-api.log because it issues a REST POST call
:command:`system snmp-comm-delete "TEST\_COMMUNITY1"` - is logged to sysinv-api.log because it issues a REST DELETE call
:command:`system snmp-comm-delete "TEST_COMMUNITY1"` - is logged to sysinv-api.log because it issues a REST DELETE call
- If the :command:`sysinv` command only issues a REST GET call, it is not logged.

View File

@ -86,12 +86,12 @@ configuration is required in order to use :command:`helm`.
.. note::
In order for your remote host to trust the certificate used by
the |prod-long| K8S API, you must ensure that the
``k8s\_root\_ca\_cert`` specified at install time is a trusted
``k8s_root_ca_cert`` specified at install time is a trusted
|CA| certificate by your host. Follow the instructions for adding
a trusted |CA| certificate for the operating system distribution
of your particular host.
If you did not specify a ``k8s\_root\_ca\_cert`` at install
If you did not specify a ``k8s_root_ca_cert`` at install
time, then specify ``--insecure-skip-tls-verify``, as shown below.
The following example configures the default ~/.kube/config. See the

View File

@ -85,7 +85,7 @@ procedure.
.. _configuring-an-external-netapp-deployment-as-the-storage-backend-mod-localhost:
#. Configure Netapps configurable parameters and run the provided
install\_netapp\_backend.yml ansible playbook to enable connectivity to
install_netapp_backend.yml ansible playbook to enable connectivity to
Netapp as a storage backend for |prod|.
#. Provide Netapp backend configurable parameters in an overrides yaml
@ -98,11 +98,11 @@ procedure.
The following parameters are mandatory:
``ansible\_become\_pass``
``ansible_become_pass``
Provide the admin password.
``netapp\_backends``
``name``
``netapp_backends``
**name**
A name for the storage class.
``provisioner``
@ -136,18 +136,18 @@ procedure.
The following parameters are optional:
``trident\_setup\_dir``
``trident_setup_dir``
Set a staging directory for generated configuration files. The
default is /tmp/trident.
``trident\_namespace``
``trident_namespace``
Set this option to use an alternate Kubernetes namespace.
``trident\_rest\_api\_port``
``trident_rest_api_port``
Use an alternate port for the Trident REST API. The default is
8000.
``trident\_install\_extra\_params``
``trident_install_extra_params``
Add extra space-separated parameters when installing trident.
For complete listings of available parameters, see
@ -190,8 +190,8 @@ procedure.
username: "admin"
password: "secret"
This file is sectioned into ``netapp\_k8s\_storageclass``,
``netapp\_k8s\_snapshotstorageclasses``, and ``netapp\_backends``.
This file is sectioned into ``netapp_k8s_storageclass``,
``netapp_k8s_snapshotstorageclasses``, and ``netapp_backends``.
You can add multiple backends and/or storage classes.
.. note::

View File

@ -18,7 +18,7 @@ complement on a host.
In the Horizon Web interface and in |CLI| output, both identifications are
shown. For example, the output of the :command:`system host-disk-show`
command includes both the **device\_node** and the **device\_path**.
command includes both the **device_node** and the **device_path**.
.. code-block:: none

View File

@ -24,7 +24,7 @@ This procedure uses standard Helm mechanisms to install a second
#. Capture a list of monitors.
This will be stored in the environment variable ``<MON\_LIST>`` and
This will be stored in the environment variable ``<MON_LIST>`` and
used in the following step.
.. code-block:: none

View File

@ -24,7 +24,7 @@ You can change the space allotted for the Ceph monitor, if required.
~(keystone_admin)]$ system ceph-mon-modify <controller> ceph_mon_gib=<size>
where ``<partition\_size>`` is the size in GiB to use for the Ceph monitor.
where ``<partition_size>`` is the size in GiB to use for the Ceph monitor.
The value must be between 21 and 40 GiB.
.. code-block:: none
@ -140,10 +140,10 @@ The following are optional arguments:
For a Ceph backend, this is a user-assigned name for the backend. The
default is **ceph-store** for a Ceph backend.
**-t,** ``--tier\_uuid``
**-t,** ``--tier_uuid``
For a Ceph backend, is the UUID of a storage tier to back.
**-c,** ``--ceph\_conf``
**-c,** ``--ceph_conf``
Location of the Ceph configuration file used for provisioning an
external backend.
@ -152,7 +152,7 @@ The following are optional arguments:
reversible.
``--ceph-mon-gib``
For a Ceph backend, this is the space in gibibytes allotted for the
For a Ceph backend, this is the space in GB allotted for the
Ceph monitor.
.. note::