diff --git a/doc/source/storage/kubernetes/install-rook-ceph-a7926a1f9b70.rst b/doc/source/storage/kubernetes/install-rook-ceph-a7926a1f9b70.rst index bdea3abf2..20334ca20 100644 --- a/doc/source/storage/kubernetes/install-rook-ceph-a7926a1f9b70.rst +++ b/doc/source/storage/kubernetes/install-rook-ceph-a7926a1f9b70.rst @@ -11,7 +11,7 @@ Install Rook Ceph Rook Ceph in an orchestrator providing a containerized solution for Ceph Storage with a specialized Kubernetes Operator to automate the management of -the cluster. It is an alternative solution to the bare metal Ceph Storage. See +the cluster. It is an alternative solution to the bare metal Ceph storage. See https://rook.io/docs/rook/latest-release/Getting-Started/intro/ for more details. @@ -19,8 +19,8 @@ details. Before configuring the deployment model and services. -- Certify that there is no no ceph-store storage backend configured on the - system: +- Ensure that there is no ceph-store storage backend configured on the + system. .. code-block:: none @@ -28,21 +28,21 @@ Before configuring the deployment model and services. - Create a storage backend for Rook Ceph, choose your deployment model (controller, dedicated, open), and the desired services (block or ecblock, - filesystem, object): + filesystem, object). .. code-block:: none ~(keystone_admin)$ system storage-backend-add ceph-rook --deployment controller --confirmed -- Create a ``host-fs ceph`` for each host that will house a Rook Ceph monitor +- Create a ``host-fs ceph`` for each host that will use a Rook Ceph monitor (preferably an ODD number of hosts): .. code-block:: none ~(keystone_admin)$ system host-fs-add ceph= -- For DX platforms, adding a floating monitor is recommended. To add a - floating monitor, the inactive controller should be locked: +- It is recommended to use a |AIO-DX| platform adding a floating monitor. To add a + floating monitor the inactive controller should be locked. .. code-block:: none @@ -51,7 +51,7 @@ Before configuring the deployment model and services. - Configure |OSDs|. - - Check the uuid of the disks of the desired host that will house the + - Check the |UUID| of the disks of the desired host that will use the |OSDs|: .. code-block:: none @@ -70,16 +70,16 @@ Before configuring the deployment model and services. ~(keystone_admin)$ system host-stor-add osd -For more details om deployment models and services see +For more details on deployment models and services see :ref:`deployment-models-for-rook-ceph-b855bd0108cf`. .. rubric:: |proc| -After configuring environment according to the chosen deployment model -correctly, Rook Ceph will install automatically. +After configuring the environment based on the selected deployment model, +Rook Ceph will be installed automatically. -Check the health of the cluster after some minutes after application applied -using any ceph commands, for example :command:`ceph status`. +Check the health of the cluster after a few minutes after the application is applied +using any ceph command, for example :command:`ceph status`. .. code-block:: none @@ -104,9 +104,9 @@ using any ceph commands, for example :command:`ceph status`. usage: 3.8 GiB used, 5.7 TiB / 5.7 TiB avail pgs: 81 active+clean -Check if the cluster contains all the desired elements. All pods should be -running or completed to the cluster to be considered healthy. You can see the -Rook Ceph pods with: +Check if the cluster contains all the required elements. All pods should be +running or completed on the cluster to be considered healthy. Use the following command +to check the Rook Ceph pods on the cluster. .. code-block:: none @@ -136,26 +136,25 @@ Rook Ceph pods with: stx-ceph-osd-audit-28553058-ms92w 0/1 Completed 0 2m5s -Additional Features and Procedures ----------------------------------- +Additional Enhancements +----------------------- -Add New OSDs on a Running Cluster +Add new OSDs on a running cluster ********************************* To add new |OSDs| to the cluster, add the new |OSD| to the platform and -re-apply the application. +reapply the application. .. code-block:: none ~(keystone_admin)$ system host-stor-add ~(keystone_admin)$ system application-apply rook-ceph - -Add New Monitor on a Running Cluster -************************************ +Add a new monitor on a running cluster +************************************** To add a new monitor to the cluster, add the ``host-fs`` to the desired host -and re-apply the application. +and reapply the application. .. code-block:: none @@ -163,58 +162,84 @@ and re-apply the application. ~(keystone_admin)$ system application-apply rook-ceph -Enable Ceph Dashboard -********************* +Enable the Ceph Dashboard +************************* -To enable Ceph Dashboard a Helm override must be provided before the -application apply. You should provide a password coded in base64. +To enable the Ceph dashboard a Helm override must be provided to the +application. Provide a password coded in base64. -Create the override file. +.. rubric:: |proc| -.. code-block:: none +#. Create the override file. - $ openssl base64 -e <<< "my_dashboard_passwd" - bXlfZGFzaGJvYXJkX3Bhc3N3ZAo= + .. code-block:: none - $ cat << EOF >> dashboard-override.yaml - cephClusterSpec: - dashboard: - enabled: true - password: "bXlfZGFzaGJvYXJkX3Bhc3N3ZAo=" - EOF + $ openssl base64 -e <<< "my_dashboard_passwd" + bXlfZGFzaGJvYXJkX3Bhc3N3ZAo= -Check Rook Ceph Pods + $ cat << EOF >> dashboard-override.yaml + cephClusterSpec: + dashboard: + enabled: true + password: "bXlfZGFzaGJvYXJkX3Bhc3N3ZAo=" + EOF + +#. Update the Helm chart with the created user-override. + + .. code-block:: none + + ~(keystone_admin)$ system helm-override-update --values dashboard-override.yaml rook-ceph rook-ceph-cluster rook-ceph + +----------------+-------------------+ + | Property | Value | + +----------------+-------------------+ + | name | rook-ceph-cluster | + | namespace | rook-ceph | + | user_overrides | cephClusterSpec: | + | | dashboard: | + | | enabled: true | + | | | + +----------------+-------------------+ + +#. Apply/reapply the Rook Ceph application. + + .. code-block:: none + + ~(keystone_admin)$ system application-apply rook-ceph + +You can access the dashboard using the following address: ``https://:30443``. + +Check Rook Ceph pods ******************** -You can check the pods of the storage cluster running the following command: +You can check the pods of the storage cluster using the following command: .. code-block:: none kubectl get pod -n rook-ceph +Installation on |AIO-SX| deployments +------------------------------------ -Installation on Simplex with controller model, 1 monitor, installing manually, services: block and cephfs ---------------------------------------------------------------------------------------------------------- +For example, you can manually install a controller model, a monitor and some services (block and cephfs) +on |AIO-SX| deployments. -In this configuration, you can add monitors and |OSDs| on the Simplex node. +In this configuration, you can add monitors and |OSDs| on the |AIO-SX| node. #. On a system with no bare metal Ceph storage backend on it, add a ceph-rook - storage back end. Use block (RBD), cephfs (default option, no need to - specify with arguments). + storage backend using block (RBD), cephfs (default option). .. code-block:: none $ system storage-backend-add ceph-rook --deployment controller --confirmed -#. Add the ``host-fs ceph`` on controller, the ``host-fs ceph`` is configured - with 20 GB. +#. Add the ``host-fs ceph`` on the controller, the ``host-fs ceph`` is configured + with 20GB. .. code-block:: none $ system host-fs-add controller-0 ceph=20 -#. To add |OSDs|, get the |UUID| of each disk to feed the - :command:`host-stor-add` command: +#. To add |OSDs|, get the |UUID| of each disk and run the :command:`host-stor-add` command. .. code-block:: none @@ -234,8 +259,7 @@ In this configuration, you can add monitors and |OSDs| on the Simplex node. | | | | | | | | | | +--------------------------------------+-----------+---------+---------+-------+------------+--------------+--------------------+--------------------------------------------+ - -#. Add all the desired disks as |OSDs|: +#. Add all the desired disks as |OSDs|. .. code-block:: none @@ -281,8 +305,8 @@ In this configuration, you can add monitors and |OSDs| on the Simplex node. | updated_at | None | +------------------+--------------------------------------------------+ -#. Check the progress of the app. With a valid configuration of ``host-fs`` - and |OSDs|, the app will apply automatically. +#. Check the progress of the application. With a valid configuration of ``host-fs`` + and |OSDs| the application will be applied automatically. .. code-block:: none @@ -290,8 +314,7 @@ In this configuration, you can add monitors and |OSDs| on the Simplex node. #or $ system application-list -#. After the app is applied the pod list of the namespace rook-ceph should - look like this: +#. After applying the application the pod list of the namespace Rook Ceph is as follows: .. code-block:: none @@ -318,33 +341,35 @@ In this configuration, you can add monitors and |OSDs| on the Simplex node. stx-ceph-osd-audit-28553058-ms92w 0/1 Completed 0 2m5s -Installation on Duplex with controller model, 3 monitors, installing manually, services: block and cephfs ---------------------------------------------------------------------------------------------------------- +Installation on |AIO-DX| deployments +------------------------------------ -In this configuration, you can add monitors and OSDs on the Duplex node. +For example, you can manually install a controller model, three monitors and some services (block and cephfs) +on |AIO-DX| deployments. + +In this configuration, you can add monitors and |OSDs| on the |AIO-DX| node. #. On a system with no bare metal Ceph storage backend on it, add a ceph-rook - storage back end. Use block (RBD), cephfs (default option, no need - to specify with arguments). + storage backend using block (RBD), cephfs (default option). .. code-block:: none $ system storage-backend-add ceph-rook --deployment controller --confirmed -#. Add the ``controller-fs`` ``ceph-float`` configured with 20 GB. +#. Add the ``controller-fs`` ``ceph-float`` configured with 20GB. .. code-block:: none $ system controllerfs-add ceph-float=20 #. Add the ``host-fs ceph`` on each controller, the ``host-fs ceph`` is - configured with 20 GB. + configured with 20GB. .. code-block:: none $ system host-fs-add controller-0 ceph=20 -#. To add |OSDs|, get the |UUID| of each disk to feed the +#. To add |OSDs|, get the |UUID| of each disk and run the :command:`host-stor-add` command. .. code-block:: none @@ -381,7 +406,7 @@ In this configuration, you can add monitors and OSDs on the Duplex node. | | | | | | | | | | +--------------------------------------+-----------+---------+---------+-------+------------+--------------+-------------------+--------------------------------------------+ -#. Add all the desired disks as |OSDs|: +#. Add all the desired disks as |OSDs|. .. code-block:: none @@ -406,7 +431,6 @@ In this configuration, you can add monitors and OSDs on the Duplex node. | updated_at | None | +------------------+--------------------------------------------------+ - # system host-stor-add controller-1 #UUID $ system host-stor-add controller-1 1e36945e-e0fb-4a72-9f96-290f9bf57523 +------------------+--------------------------------------------------+ @@ -428,8 +452,8 @@ In this configuration, you can add monitors and OSDs on the Duplex node. | updated_at | None | +------------------+--------------------------------------------------+ -#. Check the progress of the app. With a valid configuration of monitors and - |OSDs|, the app will apply automatically. +#. Check the progress of the application. With a valid configuration of monitors and + |OSDs|, the app will be applied automatically. .. code-block:: none @@ -437,8 +461,8 @@ In this configuration, you can add monitors and OSDs on the Duplex node. #or $ system application-list -#. After the app is applied the pod list of the namespace ``rook-ceph`` should - look like this: +#. After applying the application the pod list of the namespace ``rook-ceph`` should + be as follows: .. code-block:: none @@ -474,26 +498,26 @@ In this configuration, you can add monitors and OSDs on the Duplex node. rook-ceph-tools-84659bcd67-r8qbp 1/1 Running 0 22m stx-ceph-manager-689997b4f4-hk6gh 1/1 Running 0 22m +Installation on Standard deployments +------------------------------------ -Installation on Standard with dedicated model, 5 monitors, services: ecblock and cephfs ---------------------------------------------------------------------------------------- +For example, you can install on standard deployments with a dedicated model, five monitors +and services (ecblock and cephfs) -In this configuration, you can add monitors on 5 hosts and, to fit this -deployment in the dedicated model, |OSDs| will be added on workers only. -Compute-1 and Compute-2 were chosen to house the cluster |OSDs|. +In this configuration, you can add monitors on 5 hosts and fit this +deployment in a dedicated model, and |OSDs| will be added on workers only. +You can choose compute-1 and compute-2 hosts to keep the cluster |OSDs|. - -#. On a system with no bare metal Ceph storage backend on it, add a ceph-rook - storage back end. To fit in the dedicated model, the |OSDs| must be placed - on dedicated workers only. We will use ``ecblock`` instead of |RBD| and - cephfs. +#. On a system with no bare metal Ceph storage backend on it, add a Ceph Rook + storage backend using cephfs and ecblock. To fit in the dedicated model, the |OSDs| must be placed + on dedicated workers only. .. code-block:: none $ system storage-backend-add ceph-rook --deployment dedicated --confirmed --services ecblock,filesystem -#. Add all the ``host-fs`` on the nodes that will house mon, mgr and mds. In - this particular case, 5 hosts will have the ``host-fs ceph`` configured. +#. Add all the ``host-fs`` on the nodes that will keep ``mon``, ``mgr`` and ``mds``. In + this case, 5 hosts will have the ``host-fs ceph`` configured. .. code-block:: none @@ -503,8 +527,7 @@ Compute-1 and Compute-2 were chosen to house the cluster |OSDs|. $ system host-fs-add compute-1 ceph=20 $ system host-fs-add compute-2 ceph=20 -#. To add |OSDs| get the |UUID| of each disk to feed the - :command:`host-stor-add` command. +#. To add |OSDs| get the |UUID| of each disk run the :command:`host-stor-add` command. .. code-block:: none @@ -541,8 +564,8 @@ Compute-1 and Compute-2 were chosen to house the cluster |OSDs|. +--------------------------------------+-----------+---------+---------+-------+------------+--------------+-------------------+--------------------------------------------+ -#. Add all the desired disks as |OSDs|, here for sake of simplicity only one - |OSD| on compute-1 and compute-2 will be added: +#. Add all the desired disks as |OSDs|, for example, only one + |OSD| on compute-1 and compute-2 will be added. .. code-block:: none @@ -588,8 +611,8 @@ Compute-1 and Compute-2 were chosen to house the cluster |OSDs|. | updated_at | None | +------------------+--------------------------------------------------+ -#. Check the progress of the app. With a valid configuration of ``host-fs`` - and |OSDs|, the app will apply automatically. +#. Check the progress of the application. With a valid configuration of ``host-fs`` + and |OSDs|, the app will be applied automatically. .. code-block:: none @@ -599,7 +622,7 @@ Compute-1 and Compute-2 were chosen to house the cluster |OSDs|. #. After the app is applied the pod list of the namespace ``rook-ceph`` should - look like this: + be as follows: .. code-block:: none