[Fix] Remove the dependency for Ceph-Utility Deployment Jobs
1) Maps and uses default ceph deployment script for deploying Ceph Components. 2) Remove redundant Apparmor folder. Change-Id: I9e0f4391b6fddb703ef1b7818e77eaf1b8a6c7f6 Signed-off-by: diwakar thyagaraj <diwakar.chitoor.thyagaraj@att.com>
This commit is contained in:
parent
f07adaae8b
commit
0f8835305b
@ -5,4 +5,6 @@ cd ${INSTALL_PATH}
|
|||||||
|
|
||||||
# Clone dependencies
|
# Clone dependencies
|
||||||
git clone https://opendev.org/openstack/openstack-helm-infra.git
|
git clone https://opendev.org/openstack/openstack-helm-infra.git
|
||||||
|
git clone https://opendev.org/openstack/openstack-helm.git
|
||||||
|
# Install Packages
|
||||||
bash -c "./openstack-helm-infra/tools/deployment/common/000-install-packages.sh"
|
bash -c "./openstack-helm-infra/tools/deployment/common/000-install-packages.sh"
|
@ -1,221 +1,16 @@
|
|||||||
#!/bin/bash
|
#!/bin/bash
|
||||||
set -xe
|
|
||||||
namespace="utility"
|
|
||||||
CURRENT_DIR="$(pwd)"
|
CURRENT_DIR="$(pwd)"
|
||||||
# Create loop back devices for ceph osds.
|
: "${OSH_PATH:="../openstack-helm"}"
|
||||||
sudo df -lh
|
: "${OSH_INFRA_PATH:="../openstack-helm-infra"}"
|
||||||
sudo lsblk
|
|
||||||
sudo mkdir -p /var/lib/openstack-helm/ceph
|
cd "${OSH_PATH}"
|
||||||
sudo truncate -s 10G /var/lib/openstack-helm/ceph/ceph-osd-data-loopbackfile.img
|
bash -c "./tools/deployment/component/ceph/ceph.sh"
|
||||||
sudo truncate -s 8G /var/lib/openstack-helm/ceph/ceph-osd-db-wal-loopbackfile.img
|
|
||||||
sudo losetup /dev/loop0 /var/lib/openstack-helm/ceph/ceph-osd-data-loopbackfile.img
|
namespace="utility"
|
||||||
sudo losetup /dev/loop1 /var/lib/openstack-helm/ceph/ceph-osd-db-wal-loopbackfile.img
|
: ${OSH_EXTRA_HELM_ARGS:=""}
|
||||||
# lets check the devices
|
|
||||||
sudo df -lh
|
|
||||||
sudo lsblk
|
|
||||||
|
|
||||||
: ${OSH_INFRA_PATH:="../../openstack-helm-infra"}
|
|
||||||
cd "${OSH_INFRA_PATH}"
|
cd "${OSH_INFRA_PATH}"
|
||||||
|
#Deploy Ceph-provisioners
|
||||||
for CHART in ceph-mon ceph-client ceph-provisioners; do
|
|
||||||
make "${CHART}"
|
|
||||||
done
|
|
||||||
|
|
||||||
#NOTE: Deploy command
|
|
||||||
: ${OSH_EXTRA_HELM_ARGS:=""}
|
|
||||||
[ -s /tmp/ceph-fs-uuid.txt ] || uuidgen > /tmp/ceph-fs-uuid.txt
|
|
||||||
CEPH_FS_ID="$(cat /tmp/ceph-fs-uuid.txt)"
|
|
||||||
#NOTE(portdirect): to use RBD devices with Ubuntu kernels < 4.5 this
|
|
||||||
# should be set to 'hammer'
|
|
||||||
. /etc/os-release
|
|
||||||
if [ "x${ID}" == "xubuntu" ] && \
|
|
||||||
[ "$(uname -r | awk -F "." '{ print $2 }')" -lt "5" ]; then
|
|
||||||
CRUSH_TUNABLES=hammer
|
|
||||||
else
|
|
||||||
CRUSH_TUNABLES=null
|
|
||||||
fi
|
|
||||||
tee /tmp/ceph.yaml <<EOF
|
|
||||||
endpoints:
|
|
||||||
ceph_mon:
|
|
||||||
namespace: ceph
|
|
||||||
port:
|
|
||||||
mon:
|
|
||||||
default: 6789
|
|
||||||
ceph_mgr:
|
|
||||||
namespace: ceph
|
|
||||||
port:
|
|
||||||
mgr:
|
|
||||||
default: 7000
|
|
||||||
metrics:
|
|
||||||
default: 9283
|
|
||||||
network:
|
|
||||||
public: 172.17.0.1/16
|
|
||||||
cluster: 172.17.0.1/16
|
|
||||||
port:
|
|
||||||
mon: 6789
|
|
||||||
rgw: 8088
|
|
||||||
mgr: 7000
|
|
||||||
deployment:
|
|
||||||
storage_secrets: true
|
|
||||||
ceph: true
|
|
||||||
rbd_provisioner: true
|
|
||||||
cephfs_provisioner: true
|
|
||||||
client_secrets: false
|
|
||||||
rgw_keystone_user_and_endpoints: false
|
|
||||||
bootstrap:
|
|
||||||
enabled: true
|
|
||||||
conf:
|
|
||||||
rgw_ks:
|
|
||||||
enabled: false
|
|
||||||
ceph:
|
|
||||||
global:
|
|
||||||
fsid: ${CEPH_FS_ID}
|
|
||||||
mon_addr: :6789
|
|
||||||
osd_pool_default_size: 1
|
|
||||||
osd:
|
|
||||||
osd_crush_chooseleaf_type: 0
|
|
||||||
pool:
|
|
||||||
crush:
|
|
||||||
tunables: ${CRUSH_TUNABLES}
|
|
||||||
target:
|
|
||||||
osd: 1
|
|
||||||
pg_per_osd: 100
|
|
||||||
default:
|
|
||||||
crush_rule: same_host
|
|
||||||
spec:
|
|
||||||
# RBD pool
|
|
||||||
- name: rbd
|
|
||||||
application: rbd
|
|
||||||
replication: 1
|
|
||||||
percent_total_data: 40
|
|
||||||
# CephFS pools
|
|
||||||
- name: cephfs_metadata
|
|
||||||
application: cephfs
|
|
||||||
replication: 1
|
|
||||||
percent_total_data: 5
|
|
||||||
- name: cephfs_data
|
|
||||||
application: cephfs
|
|
||||||
replication: 1
|
|
||||||
percent_total_data: 10
|
|
||||||
# RadosGW pools
|
|
||||||
- name: .rgw.root
|
|
||||||
application: rgw
|
|
||||||
replication: 1
|
|
||||||
percent_total_data: 0.1
|
|
||||||
- name: default.rgw.control
|
|
||||||
application: rgw
|
|
||||||
replication: 1
|
|
||||||
percent_total_data: 0.1
|
|
||||||
- name: default.rgw.data.root
|
|
||||||
application: rgw
|
|
||||||
replication: 1
|
|
||||||
percent_total_data: 0.1
|
|
||||||
- name: default.rgw.gc
|
|
||||||
application: rgw
|
|
||||||
replication: 1
|
|
||||||
percent_total_data: 0.1
|
|
||||||
- name: default.rgw.log
|
|
||||||
application: rgw
|
|
||||||
replication: 1
|
|
||||||
percent_total_data: 0.1
|
|
||||||
- name: default.rgw.intent-log
|
|
||||||
application: rgw
|
|
||||||
replication: 1
|
|
||||||
percent_total_data: 0.1
|
|
||||||
- name: default.rgw.meta
|
|
||||||
application: rgw
|
|
||||||
replication: 1
|
|
||||||
percent_total_data: 0.1
|
|
||||||
- name: default.rgw.usage
|
|
||||||
application: rgw
|
|
||||||
replication: 1
|
|
||||||
percent_total_data: 0.1
|
|
||||||
- name: default.rgw.users.keys
|
|
||||||
application: rgw
|
|
||||||
replication: 1
|
|
||||||
percent_total_data: 0.1
|
|
||||||
- name: default.rgw.users.email
|
|
||||||
application: rgw
|
|
||||||
replication: 1
|
|
||||||
percent_total_data: 0.1
|
|
||||||
- name: default.rgw.users.swift
|
|
||||||
application: rgw
|
|
||||||
replication: 1
|
|
||||||
percent_total_data: 0.1
|
|
||||||
- name: default.rgw.users.uid
|
|
||||||
application: rgw
|
|
||||||
replication: 1
|
|
||||||
percent_total_data: 0.1
|
|
||||||
- name: default.rgw.buckets.extra
|
|
||||||
application: rgw
|
|
||||||
replication: 1
|
|
||||||
percent_total_data: 0.1
|
|
||||||
- name: default.rgw.buckets.index
|
|
||||||
application: rgw
|
|
||||||
replication: 1
|
|
||||||
percent_total_data: 3
|
|
||||||
- name: default.rgw.buckets.data
|
|
||||||
application: rgw
|
|
||||||
replication: 1
|
|
||||||
percent_total_data: 34.8
|
|
||||||
storage:
|
|
||||||
osd:
|
|
||||||
- data:
|
|
||||||
type: bluestore
|
|
||||||
location: /dev/loop0
|
|
||||||
block_db:
|
|
||||||
location: /dev/loop1
|
|
||||||
size: "5GB"
|
|
||||||
block_wal:
|
|
||||||
location: /dev/loop1
|
|
||||||
size: "2GB"
|
|
||||||
pod:
|
|
||||||
replicas:
|
|
||||||
mds: 1
|
|
||||||
mgr: 1
|
|
||||||
rgw: 1
|
|
||||||
jobs:
|
|
||||||
ceph_defragosds:
|
|
||||||
# Execute every 15 minutes for gates
|
|
||||||
cron: "*/15 * * * *"
|
|
||||||
history:
|
|
||||||
# Number of successful job to keep
|
|
||||||
successJob: 1
|
|
||||||
# Number of failed job to keep
|
|
||||||
failJob: 1
|
|
||||||
concurrency:
|
|
||||||
# Skip new job if previous job still active
|
|
||||||
execPolicy: Forbid
|
|
||||||
startingDeadlineSecs: 60
|
|
||||||
manifests:
|
|
||||||
cronjob_defragosds: true
|
|
||||||
job_bootstrap: false
|
|
||||||
EOF
|
|
||||||
|
|
||||||
for CHART in ceph-mon ceph-client ceph-provisioners; do
|
|
||||||
helm upgrade --install ${CHART} ./${CHART} \
|
|
||||||
--namespace=ceph \
|
|
||||||
--values=/tmp/ceph.yaml \
|
|
||||||
${OSH_INFRA_EXTRA_HELM_ARGS} \
|
|
||||||
${OSH_INFRA_EXTRA_HELM_ARGS_CEPH_DEPLOY}
|
|
||||||
done
|
|
||||||
helm upgrade --install ceph-osd ./ceph-osd \
|
|
||||||
--namespace=ceph \
|
|
||||||
--values=/tmp/ceph.yaml
|
|
||||||
|
|
||||||
#NOTE: Wait for deploy
|
|
||||||
./tools/deployment/common/wait-for-pods.sh ceph
|
|
||||||
|
|
||||||
#NOTE: Validate deploy
|
|
||||||
MON_POD=$(kubectl get pods \
|
|
||||||
--namespace=ceph \
|
|
||||||
--selector="application=ceph" \
|
|
||||||
--selector="component=mon" \
|
|
||||||
--no-headers | awk '{ print $1; exit }')
|
|
||||||
kubectl exec -n ceph ${MON_POD} -- ceph -s
|
|
||||||
|
|
||||||
#NOTE: Deploy command
|
|
||||||
: ${OSH_EXTRA_HELM_ARGS:=""}
|
|
||||||
tee /tmp/ceph-utility-config.yaml <<EOF
|
tee /tmp/ceph-utility-config.yaml <<EOF
|
||||||
endpoints:
|
endpoints:
|
||||||
identity:
|
identity:
|
||||||
@ -242,12 +37,12 @@ conf:
|
|||||||
EOF
|
EOF
|
||||||
|
|
||||||
helm upgrade --install ceph-utility-config ./ceph-provisioners \
|
helm upgrade --install ceph-utility-config ./ceph-provisioners \
|
||||||
--namespace=utility \
|
--namespace=$namespace \
|
||||||
--values=/tmp/ceph-utility-config.yaml \
|
--values=/tmp/ceph-utility-config.yaml \
|
||||||
${OSH_EXTRA_HELM_ARGS} \
|
${OSH_EXTRA_HELM_ARGS} \
|
||||||
${OSH_EXTRA_HELM_ARGS_CEPH_NS_ACTIVATE}
|
${OSH_EXTRA_HELM_ARGS_CEPH_NS_ACTIVATE}
|
||||||
|
|
||||||
#Deploy Ceph-Utility
|
# Deploy Ceph-Utility
|
||||||
cd ${CURRENT_DIR}
|
cd ${CURRENT_DIR}
|
||||||
helm dependency update charts/ceph-utility
|
helm dependency update charts/ceph-utility
|
||||||
helm upgrade --install ceph-utility ./charts/ceph-utility --namespace=$namespace
|
helm upgrade --install ceph-utility ./charts/ceph-utility --namespace=$namespace
|
||||||
|
@ -1,2 +0,0 @@
|
|||||||
#!/bin/bash
|
|
||||||
bash -c "../openstack-helm-infra/tools/deployment/common/001-setup-apparmor-profiles.sh"
|
|
Loading…
Reference in New Issue
Block a user