Rework ceph-guide bluestore/filestore sections
Create bluestore/filestore subsections in ceph-guide and add note about using bluestore by default - for better readability. Change-Id: Ib51706838ecf35ac274053b916b3a17f0de90758
This commit is contained in:
parent
8219542f89
commit
55455cdde2
@ -4,19 +4,21 @@
|
||||
Ceph in Kolla
|
||||
=============
|
||||
|
||||
The out-of-the-box Ceph deployment requires 3 hosts with at least one block
|
||||
device on each host that can be dedicated for sole use by Ceph. However, with
|
||||
tweaks to the Ceph cluster you can deploy a **healthy** cluster with a single
|
||||
host and a single block device.
|
||||
.. note::
|
||||
The out-of-the-box Ceph deployment requires 3 hosts with at least one block
|
||||
device on each host that can be dedicated for sole use by Ceph.
|
||||
|
||||
However, with tweaks to the Ceph cluster you can deploy a **healthy** cluster
|
||||
with a single host and a single block device.
|
||||
|
||||
Requirements
|
||||
~~~~~~~~~~~~
|
||||
------------
|
||||
|
||||
* A minimum of 3 hosts for a vanilla deploy
|
||||
* A minimum of 1 block device per host
|
||||
|
||||
Preparation
|
||||
~~~~~~~~~~~
|
||||
-----------
|
||||
|
||||
To prepare a disk for use as a
|
||||
`Ceph OSD <http://docs.ceph.com/docs/master/man/8/ceph-osd/>`_ you must add a
|
||||
@ -24,6 +26,14 @@ special partition label to the disk. This partition label is how Kolla detects
|
||||
the disks to format and bootstrap. Any disk with a matching partition label
|
||||
will be reformatted so use caution.
|
||||
|
||||
Filestore
|
||||
~~~~~~~~~
|
||||
|
||||
.. note::
|
||||
|
||||
From Rocky release - kolla-ansible by default creates Bluestore OSDs.
|
||||
Please see Configuration section to change that behaviour.
|
||||
|
||||
To prepare a filestore OSD as a storage drive, execute the following
|
||||
operations:
|
||||
|
||||
@ -53,6 +63,9 @@ usage with Kolla.
|
||||
|
||||
.. end
|
||||
|
||||
Bluestore
|
||||
~~~~~~~~~
|
||||
|
||||
To prepare a bluestore OSD partition, execute the following operations:
|
||||
|
||||
.. code-block:: console
|
||||
@ -158,7 +171,7 @@ To prepare the journal external drive execute the following command:
|
||||
|
||||
|
||||
Configuration
|
||||
~~~~~~~~~~~~~
|
||||
-------------
|
||||
|
||||
Edit the ``[storage]`` group in the inventory which contains the hostname
|
||||
of the hosts that have the block devices you have prepped as shown above.
|
||||
@ -250,7 +263,7 @@ For more details, see `NETWORK CONFIGURATION REFERENCE
|
||||
of Ceph Documentation.
|
||||
|
||||
Deployment
|
||||
~~~~~~~~~~
|
||||
----------
|
||||
|
||||
Finally deploy the Ceph-enabled OpenStack:
|
||||
|
||||
@ -260,8 +273,8 @@ Finally deploy the Ceph-enabled OpenStack:
|
||||
|
||||
.. end
|
||||
|
||||
Using a Cache Tiering
|
||||
~~~~~~~~~~~~~~~~~~~~~
|
||||
Using Cache Tiering
|
||||
-------------------
|
||||
|
||||
An optional `cache tiering <http://docs.ceph.com/docs/jewel/rados/operations/cache-tiering/>`_
|
||||
can be deployed by formatting at least one cache device and enabling cache.
|
||||
@ -304,7 +317,7 @@ After this run the playbooks as you normally would, for example:
|
||||
.. end
|
||||
|
||||
Setting up an Erasure Coded Pool
|
||||
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||
--------------------------------
|
||||
|
||||
`Erasure code <http://docs.ceph.com/docs/jewel/rados/operations/erasure-code/>`_
|
||||
is the new big thing from Ceph. Kolla has the ability to setup your Ceph pools
|
||||
@ -328,7 +341,7 @@ To enable erasure coded pools add the following options to your
|
||||
.. end
|
||||
|
||||
Managing Ceph
|
||||
~~~~~~~~~~~~~
|
||||
-------------
|
||||
|
||||
Check the Ceph status for more diagnostic information. The sample output below
|
||||
indicates a healthy cluster:
|
||||
@ -389,10 +402,10 @@ The default pool Ceph creates is named **rbd**. It is safe to remove this pool:
|
||||
.. end
|
||||
|
||||
Troubleshooting
|
||||
~~~~~~~~~~~~~~~
|
||||
---------------
|
||||
|
||||
Deploy fails with 'Fetching Ceph keyrings ... No JSON object could be decoded'
|
||||
------------------------------------------------------------------------------
|
||||
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||
|
||||
If an initial deploy of Ceph fails, perhaps due to improper configuration or
|
||||
similar, the cluster will be partially formed and will need to be reset for a
|
||||
@ -408,7 +421,7 @@ from each Ceph monitor node:
|
||||
ceph-mon
|
||||
|
||||
Simple 3 Node Example
|
||||
~~~~~~~~~~~~~~~~~~~~~
|
||||
---------------------
|
||||
|
||||
This example will show how to deploy Ceph in a very simple setup using 3
|
||||
storage nodes. 2 of those nodes (kolla1 and kolla2) will also provide other
|
||||
@ -455,7 +468,7 @@ environment before adding the 3rd node for Ceph:
|
||||
.. end
|
||||
|
||||
Configuration
|
||||
-------------
|
||||
~~~~~~~~~~~~~
|
||||
|
||||
To prepare the 2nd disk (/dev/sdb) of each nodes for use by Ceph you will need
|
||||
to add a partition label to it as shown below:
|
||||
@ -513,7 +526,7 @@ It is now time to enable Ceph in the environment by editing the
|
||||
.. end
|
||||
|
||||
Deployment
|
||||
----------
|
||||
~~~~~~~~~~
|
||||
|
||||
Finally deploy the Ceph-enabled configuration:
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user