
Storage configuration for Rook is currently located within the airship-core type; however, the multi-tenant type no longer inherits this configuration from the airship-core type. This change introduces a new composite, storage-cluster, to share the configuration between both types. Signed-off-by: Drew Walters <andrew.walters@att.com> Change-Id: Ib10098b38466725f00d9ec9d48058cd269544a3a
78 lines
3.5 KiB
YAML
78 lines
3.5 KiB
YAML
apiVersion: ceph.rook.io/v1
|
|
kind: CephCluster
|
|
metadata:
|
|
name: ceph
|
|
namespace: rook-ceph
|
|
spec:
|
|
dataDirHostPath: /var/lib/rook
|
|
cephVersion:
|
|
image: ceph/ceph:v15.2.11
|
|
#allowUnsupported: true
|
|
mon:
|
|
count: 3
|
|
allowMultiplePerNode: false
|
|
mgr:
|
|
count: 1
|
|
modules:
|
|
- name: pg_autoscaler
|
|
enabled: true
|
|
dashboard:
|
|
enabled: true
|
|
# If you are going to use the dashboard together with ingress-controller,
|
|
# make sure it is deployed.
|
|
ssl: true
|
|
crashCollector:
|
|
disable: false
|
|
network:
|
|
# Instead of 'host' you can enable the 'multus' network provider.
|
|
# However, the Multus network is in the EXPERIMENTAL stage.
|
|
provider: host
|
|
storage:
|
|
# Using settings below is not recommended for the production environment
|
|
useAllNodes: true # Recommended setting is `false`
|
|
useAllDevices: true # Recommended setting is `false`
|
|
# # To gain more control over you deployment, you should uncomment settings
|
|
# # listed below and setup your storage layout per node.
|
|
# # Please refer to the official rook documentation
|
|
# nodes:
|
|
# - name: changemes02
|
|
# devices:
|
|
# # You can use a list of devices (by path)
|
|
# # - name: /dev/disk/by-path/pci-0000:18:00.0-scsi-0:2:3:0
|
|
# # - name: /dev/disk/by-path/pci-0000:18:00.0-scsi-0:2:4:0
|
|
# # - name: /dev/disk/by-path/pci-0000:18:00.0-scsi-0:2:5:0
|
|
# # - name: /dev/disk/by-path/pci-0000:18:00.0-scsi-0:2:6:0
|
|
# # - name: /dev/disk/by-path/pci-0000:18:00.0-scsi-0:2:7:0
|
|
# # - name: /dev/disk/by-path/pci-0000:18:00.0-scsi-0:2:8:0
|
|
# - name: changemes04
|
|
# # Instead of enlisting available devices you can use regex
|
|
# devicePathFilter: "^/dev/disk/by-path/pci-0000:18:00.0-scsi-0:2:[3-8]:0"
|
|
# - name: changemes05
|
|
# devices:
|
|
# # Or you can use drive names (directly)
|
|
# # - name: /dev/sdc
|
|
# # - name: /dev/sdd
|
|
# # - name: /dev/sde
|
|
# # - name: /dev/sdf
|
|
# # - name: /dev/sdg
|
|
# # - name: /dev/sdh
|
|
# - name: changemes06
|
|
# # Or via regex
|
|
# deviceFilter: "^/dev/sd[c-h]"
|
|
# Also you can configure each device and/or each node. Please refer to the official rook
|
|
# documentation for the branch 1.5.x
|
|
# The section for configuring management of daemon disruptions during upgrade or fencing.
|
|
disruptionManagement:
|
|
# If true, the operator will create and manage PodDisruptionBudgets for OSD, Mon, RGW, and MDS daemons. OSD PDBs are managed dynamically
|
|
# via the strategy outlined in the [design](https://github.com/rook/rook/blob/master/design/ceph/ceph-managed-disruptionbudgets.md). The operator will
|
|
# block eviction of OSDs by default and unblock them safely when drains are detected.
|
|
managePodBudgets: true
|
|
# A duration in minutes that determines how long an entire failureDomain like `region/zone/host` will be held in `noout` (in addition to the
|
|
# default DOWN/OUT interval) when it is draining. This is only relevant when `managePodBudgets` is `true`. The default value is `30` minutes.
|
|
osdMaintenanceTimeout: 30
|
|
# A duration in minutes that the operator will wait for the placement groups to become healthy (active+clean) after a drain was completed and OSDs came back up.
|
|
# Operator will continue with the next drain if the timeout exceeds. It only works if `managePodBudgets` is `true`.
|
|
# No values or 0 means that the operator will wait until the placement groups are healthy before unblocking the next drain.
|
|
pgHealthCheckTimeout: 0
|
|
---
|