Fix ceph single node gate health
Ceph's health when running in the single node gate currently is 'HEALTH_WARN'. This PS adds an extra config option such that the ceph cluster will not attempt to perform replication when deployed on a single node, as well as introduces a mvp yaml file for ceph deployments in the gate. Change-Id: Ib2ec3345140f541c94da044ff9d77723ea3ee2bd
This commit is contained in:
parent
2b677ab1a8
commit
bc8da1a89d
@ -48,6 +48,7 @@ osd_max_object_name_len = {{ .Values.conf.ceph.config.osd.osd_max_object_name_le
|
|||||||
# crush
|
# crush
|
||||||
osd_pool_default_crush_rule = {{ .Values.conf.ceph.config.osd.osd_pool_default_crush_rule | default "0" | quote }}
|
osd_pool_default_crush_rule = {{ .Values.conf.ceph.config.osd.osd_pool_default_crush_rule | default "0" | quote }}
|
||||||
osd_crush_update_on_start = {{ .Values.conf.ceph.config.osd.osd_crush_update_on_start | default "true" | quote }}
|
osd_crush_update_on_start = {{ .Values.conf.ceph.config.osd.osd_crush_update_on_start | default "true" | quote }}
|
||||||
|
osd_crush_chooseleaf_type = {{ .Values.conf.ceph.config.osd.osd_crush_chooseleaf_type | default "1" | quote }}
|
||||||
|
|
||||||
# backend
|
# backend
|
||||||
osd_objectstore = {{ .Values.conf.ceph.config.osd.osd_objectstore | default "filestore" | quote }}
|
osd_objectstore = {{ .Values.conf.ceph.config.osd.osd_objectstore | default "filestore" | quote }}
|
||||||
|
@ -160,6 +160,7 @@ conf:
|
|||||||
#crush
|
#crush
|
||||||
osd_pool_default_crush_rule: 0
|
osd_pool_default_crush_rule: 0
|
||||||
osd_crush_update_on_start: true
|
osd_crush_update_on_start: true
|
||||||
|
osd_crush_chooseleaf_type: 1
|
||||||
#backend
|
#backend
|
||||||
osd_objectstore: filestore
|
osd_objectstore: filestore
|
||||||
#performance tuning
|
#performance tuning
|
||||||
|
@ -49,11 +49,20 @@ EOF"
|
|||||||
export osd_cluster_network=${SUBNET_RANGE}
|
export osd_cluster_network=${SUBNET_RANGE}
|
||||||
export osd_public_network=${SUBNET_RANGE}
|
export osd_public_network=${SUBNET_RANGE}
|
||||||
|
|
||||||
helm install --namespace=ceph ${WORK_DIR}/ceph --name=ceph \
|
if [ "x$INTEGRATION" == "xaio" ]; then
|
||||||
--set manifests_enabled.client_secrets=false \
|
helm install --namespace=ceph ${WORK_DIR}/ceph --name=ceph \
|
||||||
--set network.public=$osd_public_network \
|
--set manifests_enabled.client_secrets=false \
|
||||||
--set network.cluster=$osd_cluster_network \
|
--set network.public=$osd_public_network \
|
||||||
--set bootstrap.enabled=true
|
--set network.cluster=$osd_cluster_network \
|
||||||
|
--set bootstrap.enabled=true \
|
||||||
|
--values=${WORK_DIR}/tools/overrides/mvp/ceph.yaml
|
||||||
|
else
|
||||||
|
helm install --namespace=ceph ${WORK_DIR}/ceph --name=ceph \
|
||||||
|
--set manifests_enabled.client_secrets=false \
|
||||||
|
--set network.public=$osd_public_network \
|
||||||
|
--set network.cluster=$osd_cluster_network \
|
||||||
|
--set bootstrap.enabled=true
|
||||||
|
fi
|
||||||
|
|
||||||
kube_wait_for_pods ceph 600
|
kube_wait_for_pods ceph 600
|
||||||
|
|
||||||
|
21
tools/overrides/mvp/ceph.yaml
Normal file
21
tools/overrides/mvp/ceph.yaml
Normal file
@ -0,0 +1,21 @@
|
|||||||
|
# Copyright 2017 The Openstack-Helm Authors.
|
||||||
|
#
|
||||||
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
# you may not use this file except in compliance with the License.
|
||||||
|
# You may obtain a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing, software
|
||||||
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
# See the License for the specific language governing permissions and
|
||||||
|
# limitations under the License.
|
||||||
|
|
||||||
|
conf:
|
||||||
|
ceph:
|
||||||
|
config:
|
||||||
|
global:
|
||||||
|
osd_pool_default_size: 1
|
||||||
|
osd:
|
||||||
|
osd_crush_chooseleaf_type: 0
|
Loading…
Reference in New Issue
Block a user