
Fix and update links. Change-Id: I868b64a8b347d7746f857543f3a75760954ddee9 Signed-off-by: Elisamara Aoki Goncalves <elisamaraaoki.goncalves@windriver.com>
85 lines
4.1 KiB
ReStructuredText
85 lines
4.1 KiB
ReStructuredText
|
|
.. awp1552678699112
|
|
.. _replication-groups:
|
|
|
|
==================
|
|
Replication Groups
|
|
==================
|
|
|
|
The storage hosts on Ceph systems are organized into replication groups to
|
|
provide redundancy.
|
|
|
|
Each replication group contains a number of hosts, referred to as peers.
|
|
Each peer independently replicates the same data. |prod| supports a minimum
|
|
of two peers and a maximum of three peers per replication group. This
|
|
replication factor is defined when the Ceph storage backend is added.
|
|
|
|
For a system with two peers per replication group, up to four replication
|
|
groups are supported. For a system with three peers per replication group,
|
|
up to three replication groups are supported.
|
|
|
|
For best performance, |org| recommends a balanced storage capacity, in
|
|
which each peer has sufficient resources to meet the operational
|
|
requirements of the system.
|
|
|
|
A replication group is considered healthy when all its peers are available.
|
|
When only a minimum number of peers are available (as indicated by the
|
|
**min_replication** value reported for the group), the group continues to
|
|
provide storage services but without full replication, and a HEALTH_WARN
|
|
state is declared. When the number of available peers falls below the
|
|
**min_replication** value, the group no longer provides storage services,
|
|
and a HEALTH_ERR state is declared. The **min_replication** value is
|
|
always one less than the replication factor for the group.
|
|
|
|
It is not possible to lock more than one peer at a time in a replication
|
|
group.
|
|
|
|
Replication groups are created automatically. When a new storage host is
|
|
added and an incomplete replication group exists, the host is added to the
|
|
existing group. If all existing replication groups are complete, then a new
|
|
incomplete replication group is defined and the host becomes its first
|
|
member.
|
|
|
|
.. note::
|
|
Ceph relies on monitoring to detect when to switch from a primary OSD
|
|
to a replicated OSD. The Ceph parameter :command:`osd heartbeat grace` sets
|
|
the amount of time required to wait before switching OSDs when the
|
|
primary OSD is not responding. |prod| currently uses the default value
|
|
of 20 seconds. This means that a Ceph filesystem may not respond to I/O
|
|
for up to 20 seconds when a storage node or OSD goes out of service.
|
|
For more information, see the Ceph documentation:
|
|
http://docs.ceph.com/docs/master/rados/configuration/mon-osd-interaction.
|
|
|
|
Replication groups are shown on the Hosts Inventory page in association
|
|
with the storage hosts. You can also use the following CLI commands to
|
|
obtain information about replication groups:
|
|
|
|
.. code-block:: none
|
|
|
|
~(keystone_admin)]$ system cluster-list
|
|
+--------------------------------------+--------------+------+----------+------------------+
|
|
| uuid | cluster_uuid | type | name | deployment_model |
|
|
+--------------------------------------+--------------+------+----------+------------------+
|
|
| 335766eb-8564-4f4d-8665-681f73d13dfb | None | ceph | ceph_clu | controller-nodes |
|
|
| | | | ster | |
|
|
| | | | | |
|
|
+--------------------------------------+--------------+------+----------+------------------+
|
|
|
|
|
|
.. code-block:: none
|
|
|
|
~(keystone_admin)]$ system cluster-show 335766eb-968e-44fc-9ca7-907f93c772a1
|
|
|
|
+--------------------+----------------------------------------+
|
|
| Property | Value |
|
|
+--------------------+----------------------------------------+
|
|
| uuid | 335766eb-968e-44fc-9ca7-907f93c772a1 |
|
|
| cluster_uuid | None |
|
|
| type | ceph |
|
|
| name | ceph_cluster |
|
|
| replication_groups | ["group-0:['storage-0', 'storage-1']"] |
|
|
| storage_tiers | ['storage (in-use)'] |
|
|
| deployment_model | controller-nodes |
|
|
+--------------------+----------------------------------------+
|
|
|