Marshall Margenau 899384b805 Armada manifest updates for OSH
This updates the armada multinode manifest to include the new
format for waiting on releases to complete. This is now configured
to wait for a release to complete by checking resources with the
release_group label for a given chart.

This also adds the LDAP chart, which is required for the LMA
services

Change-Id: Ie23073a04dc4c55291c0314beb0d8d215005d353
2018-05-23 17:44:32 -05:00

285 lines
6.6 KiB
YAML

---
schema: armada/Chart/v1
metadata:
schema: metadata/Document/v1
name: helm-toolkit
data:
chart_name: helm-toolkit
release: helm-toolkit
namespace: helm-toolkit
values: {}
source:
type: local
location: ${OSH_INFRA_PATH}
subpath: helm-toolkit
reference: master
dependencies: []
---
schema: armada/Chart/v1
metadata:
schema: metadata/Document/v1
name: ceph-mon
data:
chart_name: ceph-mon
release: ceph-mon
namespace: ceph
wait:
timeout: 1800
labels:
release_group: osh-ceph-mon
install:
no_hooks: False
upgrade:
no_hooks: False
pre:
delete:
- type: job
labels:
application: ceph-mon
component: bootstrap
- type: job
labels:
application: ceph-mon
component: cephfs-client-key-generator
- type: job
labels:
application: ceph-mon
component: mds-keyring-generator
- type: job
labels:
application: ceph-mon
component: osd-keyring-generator
- type: job
labels:
application: ceph-mon
component: rgw-keyring-generator
- type: job
labels:
application: ceph-mon
component: mon-keyring-generator
- type: job
labels:
application: ceph-mon
component: mgr-keyring-generator
- type: job
labels:
application: ceph-mon
component: storage-keys-generator
values:
endpoints:
identity:
namespace: openstack
object_store:
namespace: ceph
ceph_mon:
namespace: ceph
network:
public: ${CEPH_NETWORK}
cluster: ${CEPH_NETWORK}
deployment:
storage_secrets: True
ceph: True
rbd_provisioner: True
cephfs_provisioner: True
client_secrets: False
rgw_keystone_user_and_endpoints: False
bootstrap:
enabled: True
conf:
ceph:
global:
#NOTE (portdirect): any uuidv4 value is acceptable here, though must
# be consistent if you wish to redploy whist maintaining data.
fsid: C1150D08-707D-4FEB-85E7-692BD9206DC3
rgw_ks:
enabled: True
pool:
crush:
#NOTE(portdirect): set to hammer if using kernel < 4.5
tunables: null
target:
# NOTE(portdirect): 5 nodes, with one osd per node
osd: 5
pg_per_osd: 100
storage:
osd:
- data:
type: directory
location: /var/lib/openstack-helm/ceph/osd/osd-one
journal:
type: directory
location: /var/lib/openstack-helm/ceph/osd/journal-one
source:
type: local
location: ${OSH_PATH}
subpath: ceph-mon
reference: master
dependencies:
- helm-toolkit
---
schema: armada/Chart/v1
metadata:
schema: metadata/Document/v1
name: ceph-osd
data:
chart_name: ceph-osd
release: ceph-osd
namespace: ceph
wait:
timeout: 1800
labels:
release_group: osh-ceph-osd
install:
no_hooks: False
upgrade:
no_hooks: False
values:
endpoints:
identity:
namespace: openstack
object_store:
namespace: ceph
ceph_mon:
namespace: ceph
network:
public: ${CEPH_NETWORK}
cluster: ${CEPH_NETWORK}
deployment:
storage_secrets: True
ceph: True
rbd_provisioner: True
cephfs_provisioner: True
client_secrets: False
rgw_keystone_user_and_endpoints: False
bootstrap:
enabled: True
conf:
ceph:
global:
#NOTE (portdirect): any uuidv4 value is acceptable here, though must
# be consistent if you wish to redploy whist maintaining data.
fsid: C1150D08-707D-4FEB-85E7-692BD9206DC3
rgw_ks:
enabled: True
pool:
crush:
#NOTE(portdirect): set to hammer if using kernel < 4.5
tunables: null
target:
# NOTE(portdirect): 5 nodes, with one osd per node
osd: 5
pg_per_osd: 100
storage:
osd:
- data:
type: directory
location: /var/lib/openstack-helm/ceph/osd/osd-one
journal:
type: directory
location: /var/lib/openstack-helm/ceph/osd/journal-one
source:
type: local
location: ${OSH_PATH}
subpath: ceph-osd
reference: master
dependencies:
- helm-toolkit
---
schema: armada/Chart/v1
metadata:
schema: metadata/Document/v1
name: ceph-client
data:
chart_name: ceph-client
release: ceph-client
namespace: ceph
wait:
timeout: 1800
labels:
release_group: osh-ceph-client
install:
no_hooks: False
upgrade:
no_hooks: False
pre:
delete:
- type: job
labels:
application: ceph-client
component: bootstrap
- type: job
labels:
application: ceph-client
component: cephfs-client-key-generator
- type: job
labels:
application: ceph-client
component: rbd-pool
values:
endpoints:
identity:
namespace: openstack
object_store:
namespace: ceph
ceph_mon:
namespace: ceph
network:
public: ${CEPH_NETWORK}
cluster: ${CEPH_NETWORK}
deployment:
storage_secrets: True
ceph: True
rbd_provisioner: True
cephfs_provisioner: True
client_secrets: False
rgw_keystone_user_and_endpoints: False
bootstrap:
enabled: True
conf:
rgw_ks:
enabled: True
pool:
crush:
#NOTE(portdirect): set to hammer if using kernel < 4.5
tunables: null
target:
# NOTE(portdirect): 5 nodes, with one osd per node
osd: 5
pg_per_osd: 100
storage:
osd:
- data:
type: directory
location: /var/lib/openstack-helm/ceph/osd/osd-one
journal:
type: directory
location: /var/lib/openstack-helm/ceph/osd/journal-one
source:
type: local
location: ${OSH_PATH}
subpath: ceph-client
reference: master
dependencies:
- helm-toolkit
---
schema: armada/ChartGroup/v1
metadata:
schema: metadata/Document/v1
name: ceph-storage
data:
description: "Ceph Storage"
sequenced: True
chart_group:
- ceph-mon
- ceph-osd
- ceph-client
---
schema: armada/Manifest/v1
metadata:
schema: metadata/Document/v1
name: armada-manifest
data:
release_prefix: osh
chart_groups:
- ceph-storage