Moved examples, resources and templates
new location is https://github.com/Mirantis/solar-resources, later will be changed to openstack one. - vagrant stuff assumes that solar-resources is cloned into /vagrant/solar-resources - adjusted docker compose file - added solar-resources to .gitignore Change-Id: If2fea99145395606e6c15c9adbc127ecff4823f9
This commit is contained in:
parent
0e808a985d
commit
4060b36fed
3
.gitignore
vendored
3
.gitignore
vendored
@ -57,3 +57,6 @@ solar/.cache
|
||||
.solar_config_override
|
||||
|
||||
.ropeproject
|
||||
|
||||
# for simplicity let's keep solar-resources there
|
||||
solar-resources
|
||||
|
@ -63,7 +63,7 @@ For now all commands should be executed from `solar-dev` machine from `/vagrant
|
||||
|
||||
Basic flow is:
|
||||
|
||||
1. Create some resources (look at `examples/openstack/openstack.py`) and connect
|
||||
1. Create some resources (look at `solar-resources/examples/openstack/openstack.py`) and connect
|
||||
them between each other, and place them on nodes.
|
||||
1. Run `solar changes stage` (this stages the changes)
|
||||
1. Run `solar changes process` (this prepares orchestrator graph, returning
|
||||
@ -288,7 +288,7 @@ riak_master_service.connect_list(
|
||||
)
|
||||
```
|
||||
|
||||
For full Riak example, please look at `examples/riak/riaks-template.py`.
|
||||
For full Riak example, please look at `solar-resources/examples/riak/riaks-template.py`.
|
||||
|
||||
Full documentation of individual functions is found in the `solar/template.py` file.
|
||||
|
||||
@ -301,4 +301,4 @@ Solar is shipped with sane defaults in `vagrant-setting.yaml_defaults`. If you n
|
||||
|
||||
* In `vagrant-setting.yaml_defaults` or `vagrant-settings.yaml` file uncomment `preprovisioned: false` line.
|
||||
* Run `vagrant up`, it will take some time because it builds image for bootstrap and IBP images.
|
||||
* Now you can run provisioning `/vagrant/examples/provisioning/provision.sh`
|
||||
* Now you can run provisioning `/vagrant/solar-resources/examples/provisioning/provision.sh`
|
||||
|
@ -14,6 +14,10 @@
|
||||
|
||||
- hosts: all
|
||||
tasks:
|
||||
# setup solar-resources
|
||||
# change to openstack/solar-resources later
|
||||
- git: repo=https://github.com/Mirantis/solar-resources.git dest=/vagrant/solar-resources update=no owner=vagrant
|
||||
|
||||
# set default config location
|
||||
- lineinfile:
|
||||
dest: /home/vagrant/.bashrc
|
||||
@ -48,8 +52,8 @@
|
||||
state: present
|
||||
|
||||
- file: path=/var/lib/solar/repositories state=directory owner=vagrant
|
||||
- file: src=/vagrant/resources dest=/var/lib/solar/repositories/resources state=link owner=vagrant
|
||||
- file: src=/vagrant/templates dest=/var/lib/solar/repositories/templates state=link owner=vagrant
|
||||
- file: src=/vagrant/solar-resources/resources dest=/var/lib/solar/repositories/resources state=link owner=vagrant
|
||||
- file: src=/vagrant/solar-resources/templates dest=/var/lib/solar/repositories/templates state=link owner=vagrant
|
||||
|
||||
- name: Starting docker containers
|
||||
shell: docker-compose up -d chdir=/vagrant
|
||||
|
@ -11,7 +11,11 @@ resource. Examples are:
|
||||
|
||||
Resources are defined in ``meta.yaml`` file. This file is responsible for basic
|
||||
configuration of given resource. Below is an explanation what constitutes
|
||||
typical resource.
|
||||
typical resource.
|
||||
|
||||
.. TODO: change to openstack/solar-resources later
|
||||
.. note::
|
||||
You can find example resources https://github.com/Mirantis/solar-resources
|
||||
|
||||
|
||||
Basic resource structure
|
||||
@ -40,7 +44,7 @@ as below ::
|
||||
Solar currently supports following handlers:
|
||||
|
||||
* puppet - first version of puppet handler (legacy, will be deprecated soon)
|
||||
* puppetv2 - second, improved version of puppet, supporting hiera integration
|
||||
* puppetv2 - second, improved version of puppet, supporting hiera integration
|
||||
* ansible_playbook - first version of ansible handler (legacy, will be deprecated soon)
|
||||
* ansible_template - second generation of ansible implementation, includes transport support
|
||||
|
||||
|
@ -5,9 +5,7 @@ solar-celery:
|
||||
volumes:
|
||||
- /vagrant/.vagrant:/vagrant/.vagrant
|
||||
- /vagrant:/solar
|
||||
- /vagrant/templates:/vagrant/templates
|
||||
- /vagrant/resources:/vagrant/resources
|
||||
- /vagrant/library:/vagrant/library
|
||||
- /vagrant/solar-resources:/vagrant/solar-resources
|
||||
- /root/.ssh:/root/.ssh
|
||||
- ./bootstrap/playbooks/celery.yaml:/celery.yaml
|
||||
- /var/lib/solar/repositories:/var/lib/solar/repositories
|
||||
|
@ -1,13 +0,0 @@
|
||||
# Demo of the `solar_bootstrap` Resource
|
||||
|
||||
You need to instantiate Vagrant with a slave node which is unprovisioned
|
||||
(i.e. started from the `trusty64` Vagrant box).
|
||||
|
||||
You can start the boxes from the `Vagrantfile` in master directory and
|
||||
`vagrant-settings.yml` from this directory.
|
||||
|
||||
Running
|
||||
```bash
|
||||
python example-bootstrap.py deploy
|
||||
```
|
||||
will deploy full Solar env to node `solar-dev2`.
|
@ -1,89 +0,0 @@
|
||||
#!/usr/bin/env python
|
||||
|
||||
import click
|
||||
import sys
|
||||
import time
|
||||
|
||||
from solar.core import actions
|
||||
from solar.core import resource
|
||||
from solar.core import signals
|
||||
from solar.core import validation
|
||||
from solar.core.resource import composer as cr
|
||||
from solar import errors
|
||||
from solar.dblayer.model import ModelMeta
|
||||
|
||||
|
||||
@click.group()
|
||||
def main():
|
||||
pass
|
||||
|
||||
|
||||
def setup_resources():
|
||||
ModelMeta.remove_all()
|
||||
|
||||
node2 = cr.create('node2', 'resources/ro_node/', {
|
||||
'ip': '10.0.0.4',
|
||||
'ssh_key': '/vagrant/.vagrant/machines/solar-dev2/virtualbox/private_key',
|
||||
'ssh_user': 'vagrant'
|
||||
})[0]
|
||||
|
||||
solar_bootstrap2 = cr.create('solar_bootstrap2', 'resources/solar_bootstrap', {'master_ip': '10.0.0.2'})[0]
|
||||
|
||||
signals.connect(node2, solar_bootstrap2)
|
||||
|
||||
has_errors = False
|
||||
for r in locals().values():
|
||||
if not isinstance(r, resource.Resource):
|
||||
continue
|
||||
|
||||
print 'Validating {}'.format(r.name)
|
||||
errors = validation.validate_resource(r)
|
||||
if errors:
|
||||
has_errors = True
|
||||
print 'ERROR: %s: %s' % (r.name, errors)
|
||||
|
||||
if has_errors:
|
||||
sys.exit(1)
|
||||
|
||||
resources_to_run = [
|
||||
'solar_bootstrap2',
|
||||
]
|
||||
|
||||
|
||||
@click.command()
|
||||
def deploy():
|
||||
setup_resources()
|
||||
|
||||
# run
|
||||
resources = resource.load_all()
|
||||
resources = {r.name: r for r in resources}
|
||||
|
||||
for name in resources_to_run:
|
||||
try:
|
||||
actions.resource_action(resources[name], 'run')
|
||||
except errors.SolarError as e:
|
||||
print 'WARNING: %s' % str(e)
|
||||
raise
|
||||
|
||||
time.sleep(10)
|
||||
|
||||
|
||||
@click.command()
|
||||
def undeploy():
|
||||
resources = resource.load_all()
|
||||
resources = {r.name: r for r in resources}
|
||||
|
||||
for name in reversed(resources_to_run):
|
||||
try:
|
||||
actions.resource_action(resources[name], 'remove')
|
||||
except errors.SolarError as e:
|
||||
print 'WARNING: %s' % str(e)
|
||||
|
||||
ModelMeta.remove_all()
|
||||
|
||||
main.add_command(deploy)
|
||||
main.add_command(undeploy)
|
||||
|
||||
|
||||
if __name__ == '__main__':
|
||||
main()
|
@ -1,5 +0,0 @@
|
||||
# rename it to vagrant-settings.yml then Vagrantfile
|
||||
# will use values from this file
|
||||
|
||||
slaves_count: 3
|
||||
slaves_image: ubuntu/trusty64
|
@ -1,10 +0,0 @@
|
||||
This example shows how to use solar via CLI.
|
||||
|
||||
Usage
|
||||
=====
|
||||
|
||||
Run
|
||||
|
||||
`bash ./example.sh`
|
||||
|
||||
after this you can run `solar orch report last` and wait until all tasks have status SUCCESS.
|
@ -1,18 +0,0 @@
|
||||
#!/bin/bash
|
||||
set -eux
|
||||
|
||||
function deploy {
|
||||
# this two commands will clean db
|
||||
solar resource clear_all
|
||||
|
||||
solar resource create nodes templates/nodes '{"count": 1}'
|
||||
solar resource create mariadb1 /vagrant/resources/mariadb_service image=mariadb port=3306
|
||||
solar connect node1 mariadb1
|
||||
|
||||
solar changes stage
|
||||
solar changes process
|
||||
solar orch run-once last
|
||||
solar orch report last
|
||||
}
|
||||
|
||||
deploy
|
@ -1,95 +0,0 @@
|
||||
Very simple solar example two nodes + hosts file mapping
|
||||
|
||||
Run:
|
||||
|
||||
`python examples/hosts_file/hosts.py`
|
||||
|
||||
Then you can continue with standard solar things:
|
||||
|
||||
```
|
||||
solar changes stage -d
|
||||
solar changes process
|
||||
solar or run-once last
|
||||
watch -n 1 solar or report last
|
||||
```
|
||||
|
||||
Wait until all actions have state `SUCCESS`,
|
||||
after that check `/etc/hosts` files on both nodes, it will contain entries like:
|
||||
|
||||
```
|
||||
10.0.0.3 first1441705177.99
|
||||
10.0.0.4 second1441705178.0
|
||||
```
|
||||
|
||||
If you want to try out revert functionality - you can do it in a next way:
|
||||
|
||||
After you created all the stuff, print history like this:
|
||||
|
||||
`solar ch history`
|
||||
|
||||
Output:
|
||||
|
||||
```
|
||||
log task=hosts_file1.run uid=282fe919-6059-4100-affc-56a2b3992d9d
|
||||
log task=hosts_file2.run uid=774f5a49-00f1-4bae-8a77-90d1b2d54164
|
||||
log task=node1.run uid=2559f22c-5aa9-4c05-91c6-b70884190a56
|
||||
log task=node2.run uid=18f06abe-3e8d-4356-b172-128e1dded0e6
|
||||
```
|
||||
|
||||
Now you can try to revert creation of hosts_file1
|
||||
|
||||
```
|
||||
solar ch revert 282fe919-6059-4100-affc-56a2b3992d9d
|
||||
solar ch stage
|
||||
log task=hosts_file1.remove uid=1fe456c1-a847-4902-88bf-b7f2c5687d40
|
||||
solar ch process
|
||||
solar or run-once last
|
||||
watch -n 1 solar or report last
|
||||
```
|
||||
|
||||
For now this file will be simply cleaned (more cophisticated task can be added later).
|
||||
And you can create revert of your revert, which will lead to created hosts_file1
|
||||
resource and /etc/hosts with appropriate content
|
||||
|
||||
```
|
||||
solar ch revert 282fe919-6059-4100-affc-56a2b3992d9d
|
||||
solar ch stage
|
||||
log task=hosts_file1.remove uid=1fe456c1-a847-4902-88bf-b7f2c5687d40
|
||||
solar ch process
|
||||
solar changes run-once last
|
||||
watch -n 1 solar changes report last
|
||||
```
|
||||
|
||||
After this you can revert your result of your previous revert, which will
|
||||
create this file with relevant content.
|
||||
|
||||
```
|
||||
solar ch history -n 1
|
||||
log task=hosts_file1.remove uid=1fe456c1-a847-4902-88bf-b7f2c5687d40
|
||||
solar ch revert 1fe456c1-a847-4902-88bf-b7f2c5687d40
|
||||
solar ch stage
|
||||
log task=hosts_file1.run uid=493326b2-989f-4b94-a22c-0bbd0fc5e755
|
||||
solar ch process
|
||||
solar changes run-once last
|
||||
watch -n 1 solar changes report last
|
||||
```
|
||||
|
||||
How to discard pending changes ?
|
||||
|
||||
After database was populated by some example, lets say
|
||||
```
|
||||
python examples/hosts_file/hosts.py deploy
|
||||
```
|
||||
|
||||
User is able to discard all changes with
|
||||
```
|
||||
solar ch discard
|
||||
```
|
||||
|
||||
Or any particular change with
|
||||
```
|
||||
solar ch stage
|
||||
log task=hosts_file1.run uid=a5990538-c9c6-49e4-8d58-29fae9c7aaed
|
||||
solar ch discard a5990538-c9c6-49e4-8d58-29fae9c7aaed
|
||||
```
|
||||
|
@ -1,40 +0,0 @@
|
||||
#!/usr/bin/python
|
||||
import click
|
||||
import sys
|
||||
import time
|
||||
|
||||
from solar.core import signals
|
||||
from solar.core.resource import composer as cr
|
||||
from solar.dblayer.model import ModelMeta
|
||||
|
||||
|
||||
def run():
|
||||
ModelMeta.remove_all()
|
||||
|
||||
resources = cr.create('nodes', 'templates/nodes', {'count': 2})
|
||||
|
||||
node1, node2 = [x for x in resources if x.name.startswith('node')]
|
||||
hosts1, hosts2 = [x for x in resources
|
||||
if x.name.startswith('hosts_file')]
|
||||
|
||||
node1.connect(hosts1, {
|
||||
'name': 'hosts:name',
|
||||
'ip': 'hosts:ip',
|
||||
})
|
||||
|
||||
node2.connect(hosts1, {
|
||||
'name': 'hosts:name',
|
||||
'ip': 'hosts:ip',
|
||||
})
|
||||
|
||||
node1.connect(hosts2, {
|
||||
'name': 'hosts:name',
|
||||
'ip': 'hosts:ip',
|
||||
})
|
||||
|
||||
node2.connect(hosts2, {
|
||||
'name': 'hosts:name',
|
||||
'ip': 'hosts:ip',
|
||||
})
|
||||
|
||||
run()
|
@ -1,15 +0,0 @@
|
||||
Usage
|
||||
=====
|
||||
|
||||
Run it from /vagrant dir:
|
||||
|
||||
```
|
||||
solar resource clear_all
|
||||
solar resource create nodes templates/nodes '{"count": 1}'
|
||||
solar resource create librarian_example examples/librarian/librarian.yaml '{"node": "node1"}'
|
||||
|
||||
solar changes stage
|
||||
solar changes process
|
||||
solar orch run-once
|
||||
solar orch report -w 100
|
||||
```
|
@ -1,22 +0,0 @@
|
||||
id: librarian_examples
|
||||
|
||||
resources:
|
||||
- id: rabbitmq_service1
|
||||
from: resources/rabbitmq_service
|
||||
location: #{node}#
|
||||
input:
|
||||
management_port: 15672
|
||||
port: 5672
|
||||
|
||||
- id: librarian
|
||||
location: #{node}#
|
||||
from: resources/librarian
|
||||
input:
|
||||
modules:
|
||||
- rabbitmq_service1::module::NO_EVENTS
|
||||
|
||||
events:
|
||||
- type: depends_on
|
||||
parent_action: librarian.run
|
||||
state: success
|
||||
child_action: rabbitmq_service1.run
|
@ -1,30 +0,0 @@
|
||||
Current example will do following things:
|
||||
|
||||
- fetch fuel-library from github
|
||||
- use ./update_modules.sh to fetch librarian dependencies
|
||||
- generate ceph keys on a solar-dev1
|
||||
- install ceph-mon on solar-dev1 (INPROGRESS)
|
||||
- install ceph-osd on solar-dev2 (TODO)
|
||||
- imlement removal mechanism for ceph-mon/ceph-osd (TODO)
|
||||
|
||||
|
||||
To use it:
|
||||
|
||||
```
|
||||
python exaples/library_ceph/ceph.py
|
||||
solar ch stage && solar ch process
|
||||
solar or run-once last -w 120
|
||||
```
|
||||
|
||||
If it will fail you can run particular resource action, with a lot of
|
||||
debug info.
|
||||
|
||||
```
|
||||
solar res action run ceph_mon1
|
||||
```
|
||||
|
||||
To add repositories use
|
||||
|
||||
```
|
||||
solar resource create apt1 templates/mos_repos node=node1 index=1
|
||||
```
|
@ -1,73 +0,0 @@
|
||||
|
||||
from solar.core.resource import composer as cr
|
||||
from solar.dblayer.model import ModelMeta
|
||||
import yaml
|
||||
|
||||
|
||||
STORAGE = {'objects_ceph': True,
|
||||
'osd_pool_size': 2,
|
||||
'pg_num': 128}
|
||||
|
||||
KEYSTONE = {'admin_token': 'abcde'}
|
||||
|
||||
|
||||
NETWORK_SCHEMA = {
|
||||
'endpoints': {'eth1': {'IP': ['10.0.0.3/24']}},
|
||||
'roles': {'ceph/replication': 'eth1',
|
||||
'ceph/public': 'eth1'}
|
||||
}
|
||||
|
||||
NETWORK_METADATA = yaml.load("""
|
||||
solar-dev1:
|
||||
uid: '1'
|
||||
fqdn: solar-dev1
|
||||
network_roles:
|
||||
ceph/public: 10.0.0.3
|
||||
ceph/replication: 10.0.0.3
|
||||
node_roles:
|
||||
- ceph-mon
|
||||
name: solar-dev1
|
||||
|
||||
""")
|
||||
|
||||
|
||||
def deploy():
|
||||
ModelMeta.remove_all()
|
||||
resources = cr.create('nodes', 'templates/nodes', {'count': 2})
|
||||
first_node, second_node = [x for x in resources if x.name.startswith('node')]
|
||||
first_transp = next(x for x in resources if x.name.startswith('transport'))
|
||||
|
||||
library = cr.create('library1', 'resources/fuel_library', {})[0]
|
||||
first_node.connect(library)
|
||||
|
||||
keys = cr.create('ceph_key', 'resources/ceph_keys', {})[0]
|
||||
first_node.connect(keys)
|
||||
|
||||
remote_file = cr.create('ceph_key2', 'resources/remote_file',
|
||||
{'dest': '/var/lib/astute/'})[0]
|
||||
second_node.connect(remote_file)
|
||||
keys.connect(remote_file, {'ip': 'remote_ip', 'path': 'remote_path'})
|
||||
first_transp.connect(remote_file, {'transports': 'remote'})
|
||||
|
||||
|
||||
ceph_mon = cr.create('ceph_mon1', 'resources/ceph_mon',
|
||||
{'storage': STORAGE,
|
||||
'keystone': KEYSTONE,
|
||||
'network_scheme': NETWORK_SCHEMA,
|
||||
'ceph_monitor_nodes': NETWORK_METADATA,
|
||||
'ceph_primary_monitor_node': NETWORK_METADATA,
|
||||
'role': 'controller',
|
||||
})[0]
|
||||
|
||||
managed_apt = cr.create(
|
||||
'managed_apt1', 'templates/mos_repos',
|
||||
{'node': first_node.name, 'index': 0})[-1]
|
||||
|
||||
keys.connect(ceph_mon, {})
|
||||
first_node.connect(ceph_mon,
|
||||
{'ip': ['ip', 'public_vip', 'management_vip']})
|
||||
library.connect(ceph_mon, {'puppet_modules': 'puppet_modules'})
|
||||
managed_apt.connect(ceph_mon, {})
|
||||
|
||||
if __name__ == '__main__':
|
||||
deploy()
|
@ -1,25 +0,0 @@
|
||||
Bootstraping lxc containers using solar and roles from os-ansible-deployment
|
||||
|
||||
At first run:
|
||||
|
||||
`python examples/lxc/example-lxc.py deploy`
|
||||
|
||||
It will do several things:
|
||||
|
||||
* Prepare about ~10 containers on solar-dev1
|
||||
* Add linux bridge on solar-dev and solar-dev1 with uid br-int53
|
||||
* Setup vxlan tunnel for solar-dev and solar-dev1
|
||||
* Generate ssh key and inject it into containers
|
||||
|
||||
Later this containers can be used as regular nodes in solar.
|
||||
Check rabbitmq example at the end of the file.
|
||||
|
||||
To deploy everything use usual solar commands.
|
||||
```
|
||||
solar changes stage -d
|
||||
solar changes process
|
||||
solar orch run-once last
|
||||
watch -n 1 solar orch report last
|
||||
```
|
||||
|
||||
Wait until all actions have state `SUCCESS`
|
@ -1,144 +0,0 @@
|
||||
#!/usr/bin/env python
|
||||
|
||||
# To run:
|
||||
# example-lxc.py deploy
|
||||
# solar changes stage
|
||||
# solar changes process
|
||||
# solar orch run-once last
|
||||
# watch 'solar orch report last'
|
||||
|
||||
import click
|
||||
|
||||
from solar.core import signals
|
||||
from solar.core.resource import composer as cr
|
||||
|
||||
from solar.system_log import change
|
||||
from solar.cli import orch
|
||||
from solar.dblayer.model import ModelMeta
|
||||
|
||||
|
||||
@click.group()
|
||||
def main():
|
||||
pass
|
||||
|
||||
|
||||
def lxc_template(idx):
|
||||
return {
|
||||
'user': 'root',
|
||||
'mgmt_ip': '172.18.11.{}'.format(idx),
|
||||
'container_name': 'test{}'.format(idx),
|
||||
'inventory_hostname': 'test{}'.format(idx),
|
||||
'properties':
|
||||
{'container_release': 'trusty'},
|
||||
'container_networks':
|
||||
{'mgmt': {
|
||||
'address': '172.18.11.{}'.format(idx), # address for container
|
||||
'bridge': 'br-int53', # bridge to attach veth pair
|
||||
'bridge_address': '172.18.11.253/24',
|
||||
'interface': 'eth1', # interface name in container
|
||||
'netmask': '255.255.255.0',
|
||||
'type': 'veth'}}
|
||||
}
|
||||
|
||||
|
||||
@click.command()
|
||||
def deploy():
|
||||
ModelMeta.remove_all()
|
||||
|
||||
node1 = cr.create('nodes', 'templates/nodes', {})[0]
|
||||
seed = cr.create('nodes', 'templates/seed_node', {})[0]
|
||||
|
||||
ssh_key = cr.create('ssh_key1', 'resources/ssh_key', {
|
||||
'keys_dir': '/vagrant/.ssh',
|
||||
'private_key': '/vagrant/.ssh/id_rsa',
|
||||
'public_key': '/vagrant/.ssh/id_rsa.pub',
|
||||
'passphrase': '',
|
||||
})[0]
|
||||
signals.connect(seed, ssh_key)
|
||||
|
||||
cnets1 = cr.create('cnets1', 'resources/container_networks', {
|
||||
'networks':
|
||||
{'mgmt': {
|
||||
'bridge': 'br-int53',
|
||||
'bridge_address': '172.18.11.254/24'
|
||||
}}
|
||||
})[0]
|
||||
cnets2 = cr.create('cnets2', 'resources/container_networks', {
|
||||
'networks':
|
||||
{'mgmt': {
|
||||
'bridge': 'br-int53',
|
||||
'bridge_address': '172.18.11.253/24'
|
||||
}}
|
||||
})[0]
|
||||
signals.connect(seed, cnets1)
|
||||
signals.connect(node1, cnets2)
|
||||
|
||||
vxlan_mesh1 = cr.create('vxlan_mesh1', 'resources/vxlan_mesh', {
|
||||
'id': 53,
|
||||
'parent': 'eth1',
|
||||
'master': 'br-int53'
|
||||
})[0]
|
||||
vxlan_mesh2 = cr.create('vxlan_mesh2', 'resources/vxlan_mesh', {
|
||||
'id': 53,
|
||||
'parent': 'eth1',
|
||||
'master': 'br-int53'
|
||||
})[0]
|
||||
# seed node should be connected anyway, because we need to be able to ssh
|
||||
# into containers from any node
|
||||
signals.connect(seed, vxlan_mesh1)
|
||||
signals.connect(node1, vxlan_mesh2)
|
||||
|
||||
lxc_infra1 = cr.create('lxc_infra1', 'resources/lxc_host', {})[0]
|
||||
signals.connect(node1, lxc_infra1)
|
||||
|
||||
lxc_hosts = range(28, 35)
|
||||
hosts_map = {}
|
||||
for idx in lxc_hosts:
|
||||
|
||||
lxc_host_idx = cr.create(
|
||||
'lxc_host{}'.format(idx),
|
||||
'resources/lxc_container', lxc_template(idx))[0]
|
||||
hosts_map[idx] = lxc_host_idx
|
||||
|
||||
signals.connect(node1, lxc_host_idx, {
|
||||
'ip': ['ansible_ssh_host', 'physical_host'],
|
||||
})
|
||||
# this is a required to introduce depends on relationship between lxc infre
|
||||
# and lxc container
|
||||
signals.connect(lxc_infra1, lxc_host_idx, {'provides': 'requires'})
|
||||
signals.connect(cnets2, lxc_host_idx)
|
||||
signals.connect(ssh_key, lxc_host_idx, {
|
||||
'public_key': 'pub_key',
|
||||
'private_key': 'user_key'})
|
||||
|
||||
# RABBIT
|
||||
rabbitmq_service1 = cr.create('rabbitmq_service1', 'resources/rabbitmq_service/', {
|
||||
'management_port': 15672,
|
||||
'port': 5672,
|
||||
})[0]
|
||||
openstack_vhost = cr.create('openstack_vhost', 'resources/rabbitmq_vhost/', {
|
||||
'vhost_name': 'openstack'
|
||||
})[0]
|
||||
|
||||
openstack_rabbitmq_user = cr.create('openstack_rabbitmq_user', 'resources/rabbitmq_user/', {
|
||||
'user_name': 'openstack',
|
||||
'password': 'openstack_password'
|
||||
})[0]
|
||||
|
||||
signals.connect(hosts_map[28], rabbitmq_service1, {
|
||||
'mgmt_ip': 'ip',
|
||||
'user_key': 'ssh_key',
|
||||
'user': 'ssh_user'})
|
||||
signals.connect(rabbitmq_service1, openstack_vhost)
|
||||
signals.connect(rabbitmq_service1, openstack_rabbitmq_user)
|
||||
signals.connect(openstack_vhost, openstack_rabbitmq_user, {
|
||||
'vhost_name',
|
||||
})
|
||||
|
||||
print change.send_to_orchestration()
|
||||
|
||||
main.add_command(deploy)
|
||||
|
||||
|
||||
if __name__ == '__main__':
|
||||
main()
|
@ -1,15 +0,0 @@
|
||||
Deploying simple two node OpenStack env.
|
||||
|
||||
You need to run it from main solar directory. To prepare resources run:
|
||||
|
||||
`python examples/openstack/openstack.py create_all`
|
||||
|
||||
Then to start deployment:
|
||||
|
||||
`solar changes stage
|
||||
solar changes process
|
||||
solar orch run-once last`
|
||||
|
||||
To see the progress:
|
||||
|
||||
`solar orch report`
|
@ -1,873 +0,0 @@
|
||||
#!/usr/bin/env python
|
||||
|
||||
import click
|
||||
import sys
|
||||
|
||||
from solar.core import resource
|
||||
from solar.core import signals
|
||||
from solar.core import validation
|
||||
from solar.core.resource import composer as cr
|
||||
from solar import events as evapi
|
||||
from solar.dblayer.model import ModelMeta
|
||||
|
||||
PROFILE = False
|
||||
#PROFILE = True
|
||||
|
||||
|
||||
if PROFILE:
|
||||
import StringIO
|
||||
import cProfile
|
||||
import pstats
|
||||
|
||||
pr = cProfile.Profile()
|
||||
pr.enable()
|
||||
|
||||
|
||||
# TODO
|
||||
# Resource for repository OR puppet apt-module in run.pp
|
||||
# add-apt-repository cloud-archive:juno
|
||||
# To discuss: install stuff in Docker container
|
||||
|
||||
# NOTE
|
||||
# No copy of manifests, pull from upstream (implemented in the librarian resource)
|
||||
# Official puppet manifests, not fuel-library
|
||||
|
||||
|
||||
|
||||
@click.group()
|
||||
def main():
|
||||
pass
|
||||
|
||||
|
||||
def prepare_nodes(nodes_count):
|
||||
resources = cr.create('nodes', 'templates/nodes', {"count": nodes_count})
|
||||
nodes = resources.like('node')
|
||||
resources = cr.create('nodes_network', 'templates/nodes_network', {"count": nodes_count})
|
||||
nodes_sdn = resources.like('node')
|
||||
r = {}
|
||||
|
||||
for node, node_sdn in zip(nodes, nodes_sdn):
|
||||
r[node.name] = node
|
||||
r[node_sdn.name] = node_sdn
|
||||
|
||||
# LIBRARIAN
|
||||
librarian = cr.create('librarian_{}'.format(node.name), 'resources/librarian', {})[0]
|
||||
r[librarian.name] = librarian
|
||||
|
||||
node.connect(librarian, {})
|
||||
|
||||
# NETWORKING
|
||||
# TODO(bogdando) node's IPs should be populated as br-mgmt IPs, but now are hardcoded in templates
|
||||
signals.connect(node, node_sdn)
|
||||
node_sdn.connect_with_events(librarian, {'module': 'modules'}, {})
|
||||
evapi.add_dep(librarian.name, node_sdn.name, actions=('run', 'update'))
|
||||
|
||||
signals.connect(node, node_sdn)
|
||||
node_sdn.connect_with_events(librarian, {'module': 'modules'}, {})
|
||||
evapi.add_dep(librarian.name, node_sdn.name, actions=('run', 'update'))
|
||||
|
||||
return r
|
||||
|
||||
def setup_base(node, librarian):
|
||||
# MARIADB
|
||||
mariadb_service = cr.create('mariadb_service1', 'resources/mariadb_service', {
|
||||
'image': 'mariadb',
|
||||
'port': 3306
|
||||
})[0]
|
||||
|
||||
node.connect(mariadb_service)
|
||||
|
||||
# RABBIT
|
||||
rabbitmq_service = cr.create('rabbitmq_service1', 'resources/rabbitmq_service/', {
|
||||
'management_port': 15672,
|
||||
'port': 5672,
|
||||
})[0]
|
||||
openstack_vhost = cr.create('openstack_vhost', 'resources/rabbitmq_vhost/', {
|
||||
'vhost_name': 'openstack'
|
||||
})[0]
|
||||
|
||||
openstack_rabbitmq_user = cr.create('openstack_rabbitmq_user', 'resources/rabbitmq_user/', {
|
||||
'user_name': 'openstack',
|
||||
'password': 'openstack_password'
|
||||
})[0]
|
||||
|
||||
node.connect(rabbitmq_service)
|
||||
rabbitmq_service.connect_with_events(librarian, {'module': 'modules'}, {})
|
||||
evapi.add_dep(librarian.name, rabbitmq_service.name, actions=('run', 'update'))
|
||||
rabbitmq_service.connect(openstack_vhost)
|
||||
rabbitmq_service.connect(openstack_rabbitmq_user)
|
||||
openstack_vhost.connect(openstack_rabbitmq_user, {
|
||||
'vhost_name',
|
||||
})
|
||||
return {'mariadb_service': mariadb_service,
|
||||
'rabbitmq_service1': rabbitmq_service,
|
||||
'openstack_vhost': openstack_vhost,
|
||||
'openstack_rabbitmq_user': openstack_rabbitmq_user}
|
||||
|
||||
def setup_keystone(node, librarian, mariadb_service, openstack_rabbitmq_user):
|
||||
keystone_puppet = cr.create('keystone_puppet', 'resources/keystone_puppet', {})[0]
|
||||
|
||||
keystone_puppet.connect_with_events(librarian, {'module': 'modules'}, {})
|
||||
evapi.add_dep(librarian.name, keystone_puppet.name, actions=('run', 'update'))
|
||||
|
||||
evapi.add_dep(openstack_rabbitmq_user.name, keystone_puppet.name, actions=('run', 'update'))
|
||||
keystone_db = cr.create('keystone_db', 'resources/mariadb_db/', {
|
||||
'db_name': 'keystone_db',
|
||||
'login_user': 'root'
|
||||
})[0]
|
||||
keystone_db_user = cr.create('keystone_db_user', 'resources/mariadb_user/', {
|
||||
'user_name': 'keystone',
|
||||
'user_password': 'keystone',
|
||||
})[0]
|
||||
keystone_service_endpoint = cr.create('keystone_service_endpoint', 'resources/keystone_service_endpoint', {
|
||||
'endpoint_name': 'keystone',
|
||||
'adminurl': 'http://{{admin_ip}}:{{admin_port}}/v2.0',
|
||||
'internalurl': 'http://{{internal_ip}}:{{internal_port}}/v2.0',
|
||||
'publicurl': 'http://{{public_ip}}:{{public_port}}/v2.0',
|
||||
'description': 'OpenStack Identity Service',
|
||||
'type': 'identity'
|
||||
})[0]
|
||||
|
||||
admin_tenant = cr.create('admin_tenant', 'resources/keystone_tenant', {
|
||||
'tenant_name': 'admin'
|
||||
})[0]
|
||||
admin_user = cr.create('admin_user', 'resources/keystone_user', {
|
||||
'user_name': 'admin',
|
||||
'user_password': 'admin'
|
||||
})[0]
|
||||
admin_role = cr.create('admin_role', 'resources/keystone_role', {
|
||||
'role_name': 'admin'
|
||||
})[0]
|
||||
services_tenant = cr.create('services_tenant', 'resources/keystone_tenant', {
|
||||
'tenant_name': 'services'
|
||||
})[0]
|
||||
admin_role_services = cr.create('admin_role_services', 'resources/keystone_role', {
|
||||
'role_name': 'admin'
|
||||
})[0]
|
||||
|
||||
node.connect(keystone_db)
|
||||
node.connect(keystone_db_user)
|
||||
node.connect(keystone_puppet)
|
||||
mariadb_service.connect(keystone_db, {
|
||||
'port': 'login_port',
|
||||
'root_user': 'login_user',
|
||||
'root_password': 'login_password',
|
||||
'ip' : 'db_host',
|
||||
})
|
||||
keystone_db.connect(keystone_db_user, {
|
||||
'db_name',
|
||||
'login_port',
|
||||
'login_user',
|
||||
'login_password',
|
||||
'db_host'
|
||||
})
|
||||
|
||||
node.connect(keystone_service_endpoint)
|
||||
keystone_puppet.connect(keystone_service_endpoint, {
|
||||
'admin_token': 'admin_token',
|
||||
'admin_port': ['admin_port', 'keystone_admin_port'],
|
||||
'ip': ['keystone_host', 'admin_ip', 'internal_ip', 'public_ip'],
|
||||
'port': ['internal_port', 'public_port'],
|
||||
})
|
||||
|
||||
keystone_puppet.connect(admin_tenant)
|
||||
keystone_puppet.connect(admin_tenant, {
|
||||
'admin_port': 'keystone_port',
|
||||
'ip': 'keystone_host'
|
||||
})
|
||||
admin_tenant.connect(admin_user)
|
||||
admin_user.connect(admin_role)
|
||||
admin_tenant.connect(admin_role, { 'tenant_name' })
|
||||
|
||||
admin_user.connect(admin_role_services)
|
||||
services_tenant.connect(admin_role_services, { 'tenant_name' })
|
||||
|
||||
keystone_puppet.connect(services_tenant)
|
||||
keystone_puppet.connect(services_tenant, {
|
||||
'admin_port': 'keystone_port',
|
||||
'ip': 'keystone_host'
|
||||
})
|
||||
|
||||
keystone_db.connect(keystone_puppet, {
|
||||
'db_name',
|
||||
})
|
||||
keystone_db_user.connect(keystone_puppet, {
|
||||
'user_name': 'db_user',
|
||||
'user_password': 'db_password',
|
||||
})
|
||||
mariadb_service.connect(keystone_puppet, {
|
||||
'ip': 'db_host',
|
||||
'port': 'db_port',
|
||||
})
|
||||
return {'keystone_puppet': keystone_puppet,
|
||||
'keystone_db': keystone_db,
|
||||
'keystone_db_user': keystone_db_user,
|
||||
'keystone_service_endpoint': keystone_service_endpoint,
|
||||
'admin_tenant': admin_tenant,
|
||||
'admin_user': admin_user,
|
||||
'admin_role': admin_role,
|
||||
'services_tenant': services_tenant,
|
||||
'admin_role_services': admin_role_services,
|
||||
}
|
||||
|
||||
def setup_openrc(node, keystone_puppet, admin_user):
|
||||
# OPENRC
|
||||
openrc = cr.create('openrc_file', 'resources/openrc_file', {})[0]
|
||||
|
||||
node.connect(openrc)
|
||||
keystone_puppet.connect(openrc, {'ip': 'keystone_host', 'admin_port':'keystone_port'})
|
||||
admin_user.connect(openrc, {'user_name': 'user_name','user_password':'password', 'tenant_name': 'tenant'})
|
||||
return {'openrc_file' : openrc}
|
||||
|
||||
def setup_neutron(node, librarian, rabbitmq_service, openstack_rabbitmq_user, openstack_vhost):
|
||||
# NEUTRON
|
||||
# Deploy chain neutron -> (plugins) -> neutron_server -> ( agents )
|
||||
neutron_puppet = cr.create('neutron_puppet', 'resources/neutron_puppet', {
|
||||
'core_plugin': 'neutron.plugins.ml2.plugin.Ml2Plugin'
|
||||
})[0]
|
||||
|
||||
node.connect(neutron_puppet)
|
||||
|
||||
neutron_puppet.connect_with_events(librarian, {'module': 'modules'}, {})
|
||||
evapi.add_dep(librarian.name, neutron_puppet.name, actions=('run', 'update'))
|
||||
|
||||
rabbitmq_service.connect(neutron_puppet, {
|
||||
'ip': 'rabbit_host',
|
||||
'port': 'rabbit_port'
|
||||
})
|
||||
openstack_rabbitmq_user.connect(neutron_puppet, {
|
||||
'user_name': 'rabbit_user',
|
||||
'password': 'rabbit_password'})
|
||||
openstack_vhost.connect(neutron_puppet, {
|
||||
'vhost_name': 'rabbit_virtual_host'})
|
||||
return {'neutron_puppet': neutron_puppet}
|
||||
|
||||
def setup_neutron_api(node, mariadb_service, admin_user, keystone_puppet, services_tenant, neutron_puppet):
|
||||
# NEUTRON PLUGIN AND NEUTRON API (SERVER)
|
||||
neutron_plugins_ml2 = cr.create('neutron_plugins_ml2', 'resources/neutron_plugins_ml2_puppet', {})[0]
|
||||
node.connect(neutron_plugins_ml2)
|
||||
|
||||
neutron_server_puppet = cr.create('neutron_server_puppet', 'resources/neutron_server_puppet', {
|
||||
'sync_db': True,
|
||||
})[0]
|
||||
evapi.add_dep(neutron_puppet.name, neutron_server_puppet.name, actions=('run',))
|
||||
evapi.add_dep(neutron_plugins_ml2.name, neutron_server_puppet.name, actions=('run',))
|
||||
evapi.add_dep(neutron_puppet.name, neutron_plugins_ml2.name, actions=('run',))
|
||||
|
||||
neutron_db = cr.create('neutron_db', 'resources/mariadb_db/', {
|
||||
'db_name': 'neutron_db', 'login_user': 'root'})[0]
|
||||
neutron_db_user = cr.create('neutron_db_user', 'resources/mariadb_user/', {
|
||||
'user_name': 'neutron', 'user_password': 'neutron', 'login_user': 'root'})[0]
|
||||
neutron_keystone_user = cr.create('neutron_keystone_user', 'resources/keystone_user', {
|
||||
'user_name': 'neutron',
|
||||
'user_password': 'neutron'
|
||||
})[0]
|
||||
neutron_keystone_role = cr.create('neutron_keystone_role', 'resources/keystone_role', {
|
||||
'role_name': 'admin'
|
||||
})[0]
|
||||
evapi.add_dep(neutron_keystone_role.name, neutron_server_puppet.name, actions=('run',))
|
||||
neutron_keystone_service_endpoint = cr.create('neutron_keystone_service_endpoint', 'resources/keystone_service_endpoint', {
|
||||
'endpoint_name': 'neutron',
|
||||
'adminurl': 'http://{{admin_ip}}:{{admin_port}}',
|
||||
'internalurl': 'http://{{internal_ip}}:{{internal_port}}',
|
||||
'publicurl': 'http://{{public_ip}}:{{public_port}}',
|
||||
'description': 'OpenStack Network Service',
|
||||
'type': 'network'
|
||||
})[0]
|
||||
|
||||
node.connect(neutron_db)
|
||||
node.connect(neutron_db_user)
|
||||
mariadb_service.connect(neutron_db, {
|
||||
'port': 'login_port',
|
||||
'root_password': 'login_password',
|
||||
'root_user': 'login_user',
|
||||
'ip' : 'db_host'})
|
||||
mariadb_service.connect(neutron_db_user, {'port': 'login_port', 'root_password': 'login_password'})
|
||||
neutron_db.connect(neutron_db_user, {'db_name', 'db_host'})
|
||||
neutron_db_user.connect(neutron_server_puppet, {
|
||||
'user_name':'db_user',
|
||||
'db_name':'db_name',
|
||||
'user_password':'db_password',
|
||||
'db_host' : 'db_host'})
|
||||
mariadb_service.connect(neutron_server_puppet, {
|
||||
'port': 'db_port',
|
||||
'ip' : 'db_host'})
|
||||
node.connect(neutron_server_puppet)
|
||||
admin_user.connect(neutron_server_puppet, {
|
||||
'user_name': 'auth_user',
|
||||
'user_password': 'auth_password',
|
||||
'tenant_name': 'auth_tenant'
|
||||
})
|
||||
keystone_puppet.connect(neutron_server_puppet, {
|
||||
'ip': 'auth_host',
|
||||
'port': 'auth_port'
|
||||
})
|
||||
services_tenant.connect(neutron_keystone_user)
|
||||
neutron_keystone_user.connect(neutron_keystone_role)
|
||||
keystone_puppet.connect(neutron_keystone_service_endpoint, {
|
||||
'ip': ['ip', 'keystone_host'],
|
||||
'admin_port': 'keystone_admin_port',
|
||||
'admin_token': 'admin_token',
|
||||
})
|
||||
neutron_puppet.connect(neutron_keystone_service_endpoint, {
|
||||
'ip': ['admin_ip', 'internal_ip', 'public_ip'],
|
||||
'bind_port': ['admin_port', 'internal_port', 'public_port'],
|
||||
})
|
||||
return {'neutron_server_puppet': neutron_server_puppet,
|
||||
'neutron_plugins_ml2': neutron_plugins_ml2,
|
||||
'neutron_db': neutron_db,
|
||||
'neutron_db_user': neutron_db_user,
|
||||
'neutron_keystone_user': neutron_keystone_user,
|
||||
'neutron_keystone_role': neutron_keystone_role,
|
||||
'neutron_keystone_service_endpoint': neutron_keystone_service_endpoint}
|
||||
|
||||
def setup_neutron_agent(node, neutron_server_puppet):
|
||||
# NEUTRON ML2 PLUGIN & ML2-OVS AGENT WITH GRE
|
||||
neutron_agents_ml2 = cr.create('neutron_agents_ml2', 'resources/neutron_agents_ml2_ovs_puppet', {
|
||||
# TODO(bogdando) these should come from the node network resource
|
||||
'enable_tunneling': True,
|
||||
'tunnel_types': ['gre'],
|
||||
'local_ip': '10.1.0.13' # should be the IP addr of the br-mesh int.
|
||||
})[0]
|
||||
node.connect(neutron_agents_ml2)
|
||||
evapi.add_dep(neutron_server_puppet.name, neutron_agents_ml2.name, actions=('run',))
|
||||
|
||||
# NEUTRON DHCP, L3, metadata agents
|
||||
neutron_agents_dhcp = cr.create('neutron_agents_dhcp', 'resources/neutron_agents_dhcp_puppet', {})[0]
|
||||
node.connect(neutron_agents_dhcp)
|
||||
evapi.add_dep(neutron_server_puppet.name, neutron_agents_dhcp.name, actions=('run',))
|
||||
|
||||
neutron_agents_l3 = cr.create('neutron_agents_l3', 'resources/neutron_agents_l3_puppet', {
|
||||
# TODO(bogdando) these should come from the node network resource
|
||||
'metadata_port': 8775,
|
||||
'external_network_bridge': 'br-floating',
|
||||
})[0]
|
||||
node.connect(neutron_agents_l3)
|
||||
evapi.add_dep(neutron_server_puppet.name, neutron_agents_l3.name, actions=('run',))
|
||||
|
||||
neutron_agents_metadata = cr.create('neutron_agents_metadata', 'resources/neutron_agents_metadata_puppet', {
|
||||
'sh2ared_secret': 'secret',
|
||||
})[0]
|
||||
node.connect(neutron_agents_metadata)
|
||||
neutron_server_puppet.connect(neutron_agents_metadata, {
|
||||
'auth_host', 'auth_port', 'auth_password',
|
||||
'auth_tenant', 'auth_user',
|
||||
})
|
||||
return {'neutron_agents_ml2': neutron_agents_ml2,
|
||||
'neutron_agents_dhcp': neutron_agents_dhcp,
|
||||
'neutron_agents_metadata': neutron_agents_metadata}
|
||||
|
||||
def setup_neutron_compute(node, librarian, neutron_puppet, neutron_server_puppet):
|
||||
# NEUTRON FOR COMPUTE (node1)
|
||||
# Deploy chain neutron -> (plugins) -> ( agents )
|
||||
name = node.name
|
||||
neutron_puppet2 = cr.create('neutron_puppet_{}'.format(name), 'resources/neutron_puppet', {})[0]
|
||||
|
||||
neutron_puppet2.connect_with_events(librarian, {'module': 'modules'}, {})
|
||||
evapi.add_dep(librarian.name, neutron_puppet2.name, actions=('run', 'update'))
|
||||
dep = evapi.Dep(librarian.name, 'update', state='SUCESS',
|
||||
child=neutron_puppet2.name, child_action='run')
|
||||
evapi.add_event(dep)
|
||||
|
||||
node.connect(neutron_puppet2)
|
||||
neutron_puppet.connect(neutron_puppet2, {
|
||||
'rabbit_host', 'rabbit_port',
|
||||
'rabbit_user', 'rabbit_password',
|
||||
'rabbit_virtual_host',
|
||||
'package_ensure', 'core_plugin',
|
||||
})
|
||||
|
||||
# NEUTRON OVS PLUGIN & AGENT WITH GRE FOR COMPUTE (node1)
|
||||
neutron_plugins_ml22 = cr.create('neutron_plugins_ml_{}'.format(name), 'resources/neutron_plugins_ml2_puppet', {})[0]
|
||||
node.connect(neutron_plugins_ml22)
|
||||
evapi.add_dep(neutron_puppet2.name, neutron_plugins_ml22.name, actions=('run',))
|
||||
evapi.add_dep(neutron_server_puppet.name, neutron_plugins_ml22.name, actions=('run',))
|
||||
|
||||
neutron_agents_ml22 = cr.create('neutron_agents_ml_{}'.format(name), 'resources/neutron_agents_ml2_ovs_puppet', {
|
||||
# TODO(bogdando) these should come from the node network resource
|
||||
'enable_tunneling': True,
|
||||
'tunnel_types': ['gre'],
|
||||
'local_ip': '10.1.0.14' # Should be the IP addr of the br-mesh int.
|
||||
})[0]
|
||||
node.connect(neutron_agents_ml22)
|
||||
evapi.add_dep(neutron_puppet2.name, neutron_agents_ml22.name, actions=('run',))
|
||||
evapi.add_dep(neutron_server_puppet.name, neutron_agents_ml22.name, actions=('run',))
|
||||
|
||||
return {'neutron_puppet2': neutron_puppet2,
|
||||
'neutron_plugins_ml22': neutron_plugins_ml22,
|
||||
'neutron_agents_ml22': neutron_agents_ml22}
|
||||
|
||||
def setup_cinder(node, librarian, rabbitmq_service, mariadb_service, keystone_puppet, admin_user, openstack_vhost, openstack_rabbitmq_user, services_tenant):
|
||||
# CINDER
|
||||
cinder_puppet = cr.create('cinder_puppet', 'resources/cinder_puppet', {})[0]
|
||||
cinder_db = cr.create('cinder_db', 'resources/mariadb_db/', {
|
||||
'db_name': 'cinder_db', 'login_user': 'root'})[0]
|
||||
cinder_db_user = cr.create('cinder_db_user', 'resources/mariadb_user/', {
|
||||
'user_name': 'cinder', 'user_password': 'cinder', 'login_user': 'root'})[0]
|
||||
cinder_keystone_user = cr.create('cinder_keystone_user', 'resources/keystone_user', {
|
||||
'user_name': 'cinder', 'user_password': 'cinder'})[0]
|
||||
cinder_keystone_role = cr.create('cinder_keystone_role', 'resources/keystone_role', {
|
||||
'role_name': 'admin'})[0]
|
||||
cinder_keystone_service_endpoint = cr.create(
|
||||
'cinder_keystone_service_endpoint',
|
||||
'resources/keystone_service_endpoint', {
|
||||
'endpoint_name': 'cinder',
|
||||
'adminurl': 'http://{{admin_ip}}:{{admin_port}}/v2/%(tenant_id)s',
|
||||
'internalurl': 'http://{{internal_ip}}:{{internal_port}}/v2/%(tenant_id)s',
|
||||
'publicurl': 'http://{{public_ip}}:{{public_port}}/v2/%(tenant_id)s',
|
||||
'description': 'OpenStack Block Storage Service', 'type': 'volumev2'})[0]
|
||||
|
||||
node.connect(cinder_puppet)
|
||||
cinder_puppet.connect_with_events(librarian, {'module': 'modules'}, {})
|
||||
evapi.add_dep(librarian.name, cinder_puppet.name, actions=('run', 'update'))
|
||||
|
||||
node.connect(cinder_db)
|
||||
node.connect(cinder_db_user)
|
||||
rabbitmq_service.connect(cinder_puppet, {'ip': 'rabbit_host', 'port': 'rabbit_port'})
|
||||
admin_user.connect(cinder_puppet, {'user_name': 'keystone_user', 'user_password': 'keystone_password', 'tenant_name': 'keystone_tenant'}) #?
|
||||
openstack_vhost.connect(cinder_puppet, {'vhost_name': 'rabbit_virtual_host'})
|
||||
openstack_rabbitmq_user.connect(cinder_puppet, {'user_name': 'rabbit_userid', 'password': 'rabbit_password'})
|
||||
mariadb_service.connect(cinder_db, {
|
||||
'port': 'login_port',
|
||||
'root_password': 'login_password',
|
||||
'root_user': 'login_user',
|
||||
'ip' : 'db_host'})
|
||||
mariadb_service.connect(cinder_db_user, {'port': 'login_port', 'root_password': 'login_password'})
|
||||
cinder_db.connect(cinder_db_user, {'db_name', 'db_host'})
|
||||
cinder_db_user.connect(cinder_puppet, {
|
||||
'user_name':'db_user',
|
||||
'db_name':'db_name',
|
||||
'user_password':'db_password'})
|
||||
mariadb_service.connect(cinder_puppet, {
|
||||
'port': 'db_port',
|
||||
'ip': 'db_host'})
|
||||
keystone_puppet.connect(cinder_puppet, {'ip': 'keystone_host', 'admin_port': 'keystone_port'}) #or non admin port?
|
||||
services_tenant.connect(cinder_keystone_user)
|
||||
cinder_keystone_user.connect(cinder_keystone_role)
|
||||
cinder_keystone_user.connect(cinder_puppet, {'user_name': 'keystone_user', 'tenant_name': 'keystone_tenant', 'user_password': 'keystone_password'})
|
||||
mariadb_service.connect(cinder_puppet, {'ip':'ip'})
|
||||
cinder_puppet.connect(cinder_keystone_service_endpoint, {
|
||||
'ip': ['ip', 'keystone_host', 'admin_ip', 'internal_ip', 'public_ip'],
|
||||
'port': ['admin_port', 'internal_port', 'public_port'],})
|
||||
keystone_puppet.connect(cinder_keystone_service_endpoint, {
|
||||
'admin_port': 'keystone_admin_port', 'admin_token': 'admin_token'})
|
||||
|
||||
# CINDER GLANCE
|
||||
# Deploy chain: cinder_puppet -> cinder_glance -> ( cinder_api, cinder_scheduler, cinder_volume )
|
||||
cinder_glance_puppet = cr.create('cinder_glance_puppet', 'resources/cinder_glance_puppet', {})[0]
|
||||
node.connect(cinder_glance_puppet)
|
||||
evapi.add_dep(cinder_puppet.name, cinder_glance_puppet.name, actions=('run',))
|
||||
|
||||
return {'cinder_puppet': cinder_puppet,
|
||||
'cinder_db': cinder_db,
|
||||
'cinder_db_user': cinder_db_user,
|
||||
'cinder_keystone_user': cinder_keystone_user,
|
||||
'cinder_keystone_role': cinder_keystone_role,
|
||||
'cinder_keystone_service_endpoint': cinder_keystone_service_endpoint,
|
||||
'cinder_glance_puppet': cinder_glance_puppet}
|
||||
|
||||
def setup_cinder_api(node, cinder_puppet):
|
||||
# CINDER API
|
||||
cinder_api_puppet = cr.create('cinder_api_puppet', 'resources/cinder_api_puppet', {})[0]
|
||||
node.connect(cinder_api_puppet)
|
||||
cinder_puppet.connect(cinder_api_puppet, {
|
||||
'keystone_password', 'keystone_tenant', 'keystone_user'})
|
||||
cinder_puppet.connect(cinder_api_puppet, {
|
||||
'keystone_host': 'keystone_auth_host',
|
||||
'keystone_port': 'keystone_auth_port'})
|
||||
evapi.add_react(cinder_puppet.name, cinder_api_puppet.name, actions=('update',))
|
||||
return {'cinder_api_puppet': cinder_api_puppet}
|
||||
|
||||
def setup_cinder_scheduler(node, cinder_puppet):
|
||||
# CINDER SCHEDULER
|
||||
cinder_scheduler_puppet = cr.create('cinder_scheduler_puppet', 'resources/cinder_scheduler_puppet', {})[0]
|
||||
node.connect(cinder_scheduler_puppet)
|
||||
cinder_puppet.connect(cinder_scheduler_puppet)
|
||||
evapi.add_react(cinder_puppet.name, cinder_scheduler_puppet.name, actions=('update',))
|
||||
return {'cinder_scheduler_puppet': cinder_scheduler_puppet}
|
||||
|
||||
def setup_cinder_volume(node, cinder_puppet):
|
||||
# CINDER VOLUME
|
||||
cinder_volume = cr.create('cinder_volume_{}'.format(node.name), 'resources/volume_group',
|
||||
{'path': '/root/cinder.img', 'volume_name': 'cinder-volume'})[0]
|
||||
node.connect(cinder_volume)
|
||||
|
||||
cinder_volume_puppet = cr.create('cinder_volume_puppet', 'resources/cinder_volume_puppet', {})[0]
|
||||
node.connect(cinder_volume_puppet)
|
||||
cinder_puppet.connect(cinder_volume_puppet)
|
||||
evapi.add_react(cinder_puppet.name, cinder_volume_puppet.name, actions=('update',))
|
||||
cinder_volume.connect(cinder_volume_puppet, {'volume_name': 'volume_group'})
|
||||
return {'cinder_volume_puppet': cinder_volume_puppet}
|
||||
|
||||
def setup_nova(node, librarian, mariadb_service, rabbitmq_service, admin_user, openstack_vhost, services_tenant, keystone_puppet, openstack_rabbitmq_user):
|
||||
# NOVA
|
||||
nova_puppet = cr.create('nova_puppet', 'resources/nova_puppet', {})[0]
|
||||
nova_db = cr.create('nova_db', 'resources/mariadb_db/', {
|
||||
'db_name': 'nova_db',
|
||||
'login_user': 'root'})[0]
|
||||
nova_db_user = cr.create('nova_db_user', 'resources/mariadb_user/', {
|
||||
'user_name': 'nova',
|
||||
'user_password': 'nova',
|
||||
'login_user': 'root'})[0]
|
||||
nova_keystone_user = cr.create('nova_keystone_user', 'resources/keystone_user', {
|
||||
'user_name': 'nova',
|
||||
'user_password': 'nova'})[0]
|
||||
nova_keystone_role = cr.create('nova_keystone_role', 'resources/keystone_role', {
|
||||
'role_name': 'admin'})[0]
|
||||
nova_keystone_service_endpoint = cr.create('nova_keystone_service_endpoint', 'resources/keystone_service_endpoint', {
|
||||
'endpoint_name': 'nova',
|
||||
'adminurl': 'http://{{admin_ip}}:{{admin_port}}/v2/%(tenant_id)s',
|
||||
'internalurl': 'http://{{internal_ip}}:{{internal_port}}/v2/%(tenant_id)s',
|
||||
'publicurl': 'http://{{public_ip}}:{{public_port}}/v2/%(tenant_id)s',
|
||||
'description': 'OpenStack Compute Service',
|
||||
'type': 'compute'})[0]
|
||||
|
||||
node.connect(nova_puppet)
|
||||
nova_puppet.connect_with_events(librarian, {'module': 'modules'}, {})
|
||||
evapi.add_dep(librarian.name, nova_puppet.name, actions=('run', 'update'))
|
||||
|
||||
node.connect(nova_db)
|
||||
node.connect(nova_db_user)
|
||||
mariadb_service.connect(nova_db, {
|
||||
'port': 'login_port',
|
||||
'root_password': 'login_password',
|
||||
'root_user': 'login_user',
|
||||
'ip' : 'db_host'})
|
||||
mariadb_service.connect(nova_db_user, {
|
||||
'port': 'login_port',
|
||||
'root_password': 'login_password'})
|
||||
admin_user.connect(nova_puppet, {'user_name': 'keystone_user', 'user_password': 'keystone_password', 'tenant_name': 'keystone_tenant'}) #?
|
||||
openstack_vhost.connect(nova_puppet, {'vhost_name': 'rabbit_virtual_host'})
|
||||
nova_db.connect(nova_db_user, {'db_name', 'db_host'})
|
||||
services_tenant.connect(nova_keystone_user)
|
||||
nova_keystone_user.connect(nova_keystone_role)
|
||||
keystone_puppet.connect(nova_puppet, {
|
||||
'ip': 'keystone_host',
|
||||
'admin_port': 'keystone_port'})
|
||||
nova_keystone_user.connect(nova_puppet, {
|
||||
'user_name': 'keystone_user',
|
||||
'tenant_name': 'keystone_tenant',
|
||||
'user_password': 'keystone_password'})
|
||||
rabbitmq_service.connect(nova_puppet, {
|
||||
'ip': 'rabbit_host', 'port': 'rabbit_port'})
|
||||
openstack_rabbitmq_user.connect(nova_puppet, {
|
||||
'user_name': 'rabbit_userid',
|
||||
'password': 'rabbit_password'})
|
||||
keystone_puppet.connect(nova_keystone_service_endpoint, {
|
||||
'ip': 'keystone_host',
|
||||
'admin_port': 'keystone_admin_port',
|
||||
'admin_token': 'admin_token'})
|
||||
mariadb_service.connect(nova_puppet, {
|
||||
'ip':'db_host',
|
||||
'port': 'db_port'})
|
||||
nova_db_user.connect(nova_puppet, {
|
||||
'user_name':'db_user',
|
||||
'db_name':'db_name',
|
||||
'user_password':'db_password'})
|
||||
nova_puppet.connect(nova_keystone_service_endpoint, {
|
||||
'ip': ['ip', 'keystone_host', 'public_ip', 'internal_ip', 'admin_ip'],
|
||||
'port': ['admin_port', 'internal_port', 'public_port'],
|
||||
})
|
||||
return {'nova_puppet': nova_puppet,
|
||||
'nova_db': nova_db,
|
||||
'nova_db_user': nova_db_user,
|
||||
'nova_keystone_user': nova_keystone_user,
|
||||
'nova_keystone_role': nova_keystone_role,
|
||||
'nova_keystone_service_endpoint': nova_keystone_service_endpoint}
|
||||
|
||||
def setup_nova_api(node, nova_puppet, neutron_agents_metadata):
|
||||
# NOVA API
|
||||
nova_api_puppet = cr.create('nova_api_puppet', 'resources/nova_api_puppet', {})[0]
|
||||
node.connect(nova_api_puppet)
|
||||
nova_puppet.connect(nova_api_puppet, {
|
||||
'keystone_tenant': 'admin_tenant_name',
|
||||
'keystone_user': 'admin_user',
|
||||
'keystone_password': 'admin_password',
|
||||
'keystone_host': 'auth_host',
|
||||
'keystone_port': 'auth_port'})
|
||||
evapi.add_react(nova_puppet.name, nova_api_puppet.name, actions=('update',))
|
||||
nova_api_puppet.connect(neutron_agents_metadata, {'ip': 'metadata_ip'})
|
||||
return {'nova_api_puppet': nova_api_puppet}
|
||||
|
||||
def setup_nova_conductor(node, nova_puppet, nova_api_puppet):
|
||||
# NOVA CONDUCTOR
|
||||
nova_conductor_puppet = cr.create('nova_conductor_puppet', 'resources/nova_conductor_puppet', {})[0]
|
||||
node.connect(nova_conductor_puppet)
|
||||
nova_puppet.connect(nova_conductor_puppet)
|
||||
evapi.add_dep(nova_api_puppet.name, nova_conductor_puppet.name, actions=('run',))
|
||||
evapi.add_react(nova_puppet.name, nova_conductor_puppet.name, actions=('update',))
|
||||
return {'nova_conductor': nova_conductor_puppet}
|
||||
|
||||
def setup_nova_scheduler(node, nova_puppet, nova_api_puppet):
|
||||
# NOVA SCHEDULER
|
||||
# NOTE(bogdando) Generic service is used. Package and service names for Ubuntu case
|
||||
# come from https://github.com/openstack/puppet-nova/blob/5.1.0/manifests/params.pp
|
||||
nova_scheduler_puppet = cr.create('nova_scheduler_puppet', 'resources/nova_generic_service_puppet', {
|
||||
'title' : 'scheduler', 'package_name': 'nova-scheduler', 'service_name': 'nova-scheduler',
|
||||
})[0]
|
||||
node.connect(nova_scheduler_puppet)
|
||||
evapi.add_dep(nova_puppet.name, nova_scheduler_puppet.name, actions=('run',))
|
||||
evapi.add_dep(nova_api_puppet.name, nova_scheduler_puppet.name, actions=('run',))
|
||||
evapi.add_react(nova_puppet.name, nova_scheduler_puppet.name, actions=('update',))
|
||||
return {'nova_scheduler_puppet': nova_scheduler_puppet}
|
||||
|
||||
def setup_nova_compute(node, librarian, nova_puppet, nova_api_puppet, neutron_server_puppet, neutron_keystone_service_endpoint, glance_api_puppet):
|
||||
# NOVA COMPUTE
|
||||
# Deploy chain (nova, node_networking(TODO)) -> (nova_compute_libvirt, nova_neutron) -> nova_compute
|
||||
name = node.name
|
||||
nova_compute_puppet = cr.create('nova_compute_puppet_{}'.format(name), 'resources/nova_compute_puppet', {})[0]
|
||||
# TODO (bogdando) figure out how to use it for multiple glance api servers
|
||||
nova_puppet2 = cr.create('nova_puppet_{}'.format(name), 'resources/nova_puppet', {
|
||||
'glance_api_servers': '{{glance_api_servers_host}}:{{glance_api_servers_port}}'
|
||||
})[0]
|
||||
nova_puppet.connect(nova_puppet2, {
|
||||
'ensure_package', 'rabbit_host',
|
||||
'rabbit_password', 'rabbit_port', 'rabbit_userid',
|
||||
'rabbit_virtual_host', 'db_user', 'db_password',
|
||||
'db_name', 'db_host', 'keystone_password',
|
||||
'keystone_port', 'keystone_host', 'keystone_tenant',
|
||||
'keystone_user',
|
||||
})
|
||||
# TODO(bogdando): Make a connection for nova_puppet2.glance_api_servers = "glance_api_puppet.ip:glance_api_puppet.bind_port"
|
||||
node.connect(nova_puppet2)
|
||||
nova_puppet2.connect_with_events(librarian, {'module': 'modules'}, {})
|
||||
evapi.add_dep(librarian.name, nova_puppet2.name, actions=('run', 'update'))
|
||||
dep = evapi.Dep(librarian.name, 'update', state='SUCESS',
|
||||
child=nova_puppet2.name, child_action='run')
|
||||
evapi.add_event(dep)
|
||||
|
||||
node.connect(nova_compute_puppet)
|
||||
evapi.add_dep(nova_puppet2.name, nova_compute_puppet.name, actions=('run',))
|
||||
evapi.add_dep(nova_api_puppet.name, nova_compute_puppet.name, actions=('run',))
|
||||
evapi.add_react(nova_puppet2.name, nova_compute_puppet.name, actions=('run', 'update'))
|
||||
|
||||
# NOVA COMPUTE LIBVIRT, NOVA_NEUTRON
|
||||
# NOTE(bogdando): changes nova config, so should notify nova compute service
|
||||
nova_compute_libvirt_puppet = cr.create('nova_compute_libvirt_puppet_{}'.format(name), 'resources/nova_compute_libvirt_puppet', {})[0]
|
||||
node.connect(nova_compute_libvirt_puppet)
|
||||
evapi.add_dep(nova_puppet2.name, nova_compute_libvirt_puppet.name, actions=('run',))
|
||||
evapi.add_dep(nova_api_puppet.name, nova_compute_libvirt_puppet.name, actions=('run',))
|
||||
|
||||
# compute configuration for neutron, use http auth/endpoint protocols, keystone v2 auth hardcoded for the resource
|
||||
nova_neutron_puppet = cr.create('nova_neutron_puppet_{}'.format(name), 'resources/nova_neutron_puppet', {})[0]
|
||||
node.connect(nova_neutron_puppet)
|
||||
evapi.add_dep(nova_puppet2.name, nova_neutron_puppet.name, actions=('run',))
|
||||
evapi.add_dep(nova_api_puppet.name, nova_neutron_puppet.name, actions=('run',))
|
||||
neutron_server_puppet.connect(nova_neutron_puppet, {
|
||||
'auth_password': 'neutron_admin_password',
|
||||
'auth_user': 'neutron_admin_username',
|
||||
'auth_type': 'neutron_auth_strategy',
|
||||
'auth_host': 'auth_host', 'auth_port': 'auth_port',
|
||||
'auth_protocol': 'auth_protocol',
|
||||
})
|
||||
neutron_keystone_service_endpoint.connect(nova_neutron_puppet, {
|
||||
'internal_ip':'neutron_endpoint_host',
|
||||
'internal_port':'neutron_endpoint_port',
|
||||
})
|
||||
# Update glance_api_service for nova compute
|
||||
glance_api_puppet.connect(nova_puppet2, {
|
||||
'ip': 'glance_api_servers_host',
|
||||
'bind_port': 'glance_api_servers_port'
|
||||
})
|
||||
|
||||
# signals.connect(keystone_puppet, nova_network_puppet, {'ip': 'keystone_host', 'port': 'keystone_port'})
|
||||
# signals.connect(keystone_puppet, nova_keystone_service_endpoint, {'ip': 'keystone_host', 'admin_port': 'keystone_port', 'admin_token': 'admin_token'})
|
||||
# signals.connect(rabbitmq_service1, nova_network_puppet, {'ip': 'rabbitmq_host', 'port': 'rabbitmq_port'})
|
||||
return {'nova_compute_puppet': nova_compute_puppet,
|
||||
'nova_puppet2': nova_puppet2,
|
||||
'nova_compute_libvirt_puppet': nova_compute_libvirt_puppet,
|
||||
'nova_neutron_puppet': nova_neutron_puppet,
|
||||
'neutron_server_puppet': neutron_server_puppet}
|
||||
|
||||
def setup_glance_api(node, librarian, mariadb_service, admin_user, keystone_puppet, services_tenant, cinder_glance_puppet):
|
||||
# GLANCE (base and API)
|
||||
glance_api_puppet = cr.create('glance_api_puppet', 'resources/glance_puppet', {})[0]
|
||||
glance_db_user = cr.create('glance_db_user', 'resources/mariadb_user/', {
|
||||
'user_name': 'glance', 'user_password': 'glance', 'login_user': 'root'})[0]
|
||||
glance_db = cr.create('glance_db', 'resources/mariadb_db/', {
|
||||
'db_name': 'glance', 'login_user': 'root'})[0]
|
||||
glance_keystone_user = cr.create('glance_keystone_user', 'resources/keystone_user', {
|
||||
'user_name': 'glance', 'user_password': 'glance123'})[0]
|
||||
glance_keystone_role = cr.create('glance_keystone_role', 'resources/keystone_role', {
|
||||
'role_name': 'admin'})[0]
|
||||
glance_keystone_service_endpoint = cr.create(
|
||||
'glance_keystone_service_endpoint',
|
||||
'resources/keystone_service_endpoint', {
|
||||
'endpoint_name': 'glance',
|
||||
'adminurl': 'http://{{admin_ip}}:{{admin_port}}',
|
||||
'internalurl': 'http://{{internal_ip}}:{{internal_port}}',
|
||||
'publicurl': 'http://{{public_ip}}:{{public_port}}',
|
||||
'description': 'OpenStack Image Service', 'type': 'image'})[0]
|
||||
|
||||
node.connect(glance_api_puppet)
|
||||
glance_api_puppet.connect_with_events(librarian, {'module': 'modules'}, {})
|
||||
evapi.add_dep(librarian.name, glance_api_puppet.name, actions=('run', 'update'))
|
||||
|
||||
node.connect(glance_db)
|
||||
node.connect(glance_db_user)
|
||||
admin_user.connect(glance_api_puppet, {
|
||||
'user_name': 'keystone_user', 'user_password': 'keystone_password',
|
||||
'tenant_name': 'keystone_tenant'}) #?
|
||||
mariadb_service.connect(glance_db, {
|
||||
'port': 'login_port',
|
||||
'root_password': 'login_password',
|
||||
'root_user': 'login_user',
|
||||
'ip' : 'db_host'})
|
||||
mariadb_service.connect(glance_db_user, {'port': 'login_port', 'root_password': 'login_password'})
|
||||
glance_db.connect(glance_db_user, {'db_name', 'db_host'})
|
||||
glance_db_user.connect(glance_api_puppet, {
|
||||
'user_name':'db_user',
|
||||
'db_name':'db_name',
|
||||
'user_password':'db_password',
|
||||
'db_host' : 'db_host'})
|
||||
mariadb_service.connect(glance_api_puppet,{
|
||||
'port': 'db_port',
|
||||
'ip': 'db_host'})
|
||||
keystone_puppet.connect(glance_api_puppet, {'ip': 'keystone_host', 'admin_port': 'keystone_port'}) #or non admin port?
|
||||
services_tenant.connect(glance_keystone_user)
|
||||
glance_keystone_user.connect(glance_keystone_role)
|
||||
glance_keystone_user.connect(glance_api_puppet, {
|
||||
'user_name': 'keystone_user', 'tenant_name': 'keystone_tenant',
|
||||
'user_password': 'keystone_password'})
|
||||
mariadb_service.connect(glance_api_puppet, {'ip':'ip'})
|
||||
glance_api_puppet.connect(glance_keystone_service_endpoint, {
|
||||
'ip': ['ip', 'keystone_host', 'admin_ip', 'internal_ip', 'public_ip'],
|
||||
'bind_port': ['admin_port', 'internal_port', 'public_port'],})
|
||||
keystone_puppet.connect(glance_keystone_service_endpoint, {
|
||||
'admin_port': 'keystone_admin_port', 'admin_token': 'admin_token'})
|
||||
|
||||
# Update glance_api_service for cinder
|
||||
glance_api_puppet.connect(cinder_glance_puppet, {
|
||||
'ip': 'glance_api_servers_host',
|
||||
'bind_port': 'glance_api_servers_port'
|
||||
})
|
||||
return {'glance_api_puppet': glance_api_puppet,
|
||||
'glance_db_user': glance_db_user,
|
||||
'glance_db': glance_db,
|
||||
'glance_keystone_user': glance_keystone_user,
|
||||
'glance_keystone_role': glance_keystone_role,
|
||||
'glance_keystone_service_endpoint': glance_keystone_service_endpoint}
|
||||
|
||||
def setup_glance_registry(node, glance_api_puppet):
|
||||
# GLANCE REGISTRY
|
||||
glance_registry_puppet = cr.create('glance_registry_puppet', 'resources/glance_registry_puppet', {})[0]
|
||||
node.connect(glance_registry_puppet)
|
||||
glance_api_puppet.connect(glance_registry_puppet)
|
||||
evapi.add_react(glance_api_puppet.name, glance_registry_puppet.name, actions=('update',))
|
||||
# API and registry should not listen same ports
|
||||
# should not use the same log destination and a pipeline,
|
||||
# so disconnect them and restore the defaults
|
||||
signals.disconnect_receiver_by_input(glance_registry_puppet, 'bind_port')
|
||||
signals.disconnect_receiver_by_input(glance_registry_puppet, 'log_file')
|
||||
signals.disconnect_receiver_by_input(glance_registry_puppet, 'pipeline')
|
||||
glance_registry_puppet.update({
|
||||
'bind_port': 9191,
|
||||
'log_file': '/var/log/glance/registry.log',
|
||||
'pipeline': 'keystone',
|
||||
})
|
||||
return {'glance_registry_puppet': glance_registry_puppet}
|
||||
|
||||
|
||||
def validate():
|
||||
has_errors = False
|
||||
for r in locals().values():
|
||||
if not isinstance(r, resource.Resource):
|
||||
continue
|
||||
|
||||
print 'Validating {}'.format(r.name)
|
||||
errors = validation.validate_resource(r)
|
||||
if errors:
|
||||
has_errors = True
|
||||
print 'ERROR: %s: %s' % (r.name, errors)
|
||||
|
||||
if has_errors:
|
||||
sys.exit(1)
|
||||
|
||||
|
||||
def create_controller(node):
|
||||
r = {r.name: r for r in resource.load_all()}
|
||||
librarian_node = 'librarian_{}'.format(node)
|
||||
|
||||
r.update(setup_base(r[node], r[librarian_node]))
|
||||
r.update(setup_keystone(r[node], r[librarian_node],
|
||||
r['mariadb_service'], r['openstack_rabbitmq_user']))
|
||||
r.update(setup_openrc(r[node], r['keystone_puppet'], r['admin_user']))
|
||||
r.update(setup_neutron(r[node], r['librarian_{}'.format(node)], r['rabbitmq_service1'],
|
||||
r['openstack_rabbitmq_user'], r['openstack_vhost']))
|
||||
r.update(setup_neutron_api(r[node], r['mariadb_service'], r['admin_user'],
|
||||
r['keystone_puppet'], r['services_tenant'], r['neutron_puppet']))
|
||||
r.update(setup_neutron_agent(r[node], r['neutron_server_puppet']))
|
||||
r.update(setup_cinder(r[node], r['librarian_{}'.format(node)], r['rabbitmq_service1'],
|
||||
r['mariadb_service'], r['keystone_puppet'], r['admin_user'],
|
||||
r['openstack_vhost'], r['openstack_rabbitmq_user'], r['services_tenant']))
|
||||
r.update(setup_cinder_api(r[node], r['cinder_puppet']))
|
||||
r.update(setup_cinder_scheduler(r[node], r['cinder_puppet']))
|
||||
r.update(setup_cinder_volume(r[node], r['cinder_puppet']))
|
||||
r.update(setup_nova(r[node], r['librarian_{}'.format(node)], r['mariadb_service'], r['rabbitmq_service1'],
|
||||
r['admin_user'], r['openstack_vhost'], r['services_tenant'],
|
||||
r['keystone_puppet'], r['openstack_rabbitmq_user']))
|
||||
r.update(setup_nova_api(r[node], r['nova_puppet'], r['neutron_agents_metadata']))
|
||||
r.update(setup_nova_conductor(r[node], r['nova_puppet'], r['nova_api_puppet']))
|
||||
r.update(setup_nova_scheduler(r[node], r['nova_puppet'], r['nova_api_puppet']))
|
||||
r.update(setup_glance_api(r[node], r['librarian_{}'.format(node)], r['mariadb_service'], r['admin_user'],
|
||||
r['keystone_puppet'], r['services_tenant'],
|
||||
r['cinder_glance_puppet']))
|
||||
r.update(setup_glance_registry(r[node], r['glance_api_puppet']))
|
||||
return r
|
||||
|
||||
def create_compute(node):
|
||||
r = {r.name: r for r in resource.load_all()}
|
||||
librarian_node = 'librarian_{}'.format(node)
|
||||
res = {}
|
||||
res.update(setup_neutron_compute(r[node], r[librarian_node], r['neutron_puppet'], r['neutron_server_puppet']))
|
||||
res.update(setup_nova_compute(r[node], r[librarian_node], r['nova_puppet'], r['nova_api_puppet'],
|
||||
r['neutron_server_puppet'], r['neutron_keystone_service_endpoint'], r['glance_api_puppet']))
|
||||
return r
|
||||
|
||||
@click.command()
|
||||
def create_all():
|
||||
ModelMeta.remove_all()
|
||||
r = prepare_nodes(2)
|
||||
r.update(create_controller('node1'))
|
||||
r.update(create_compute('node2'))
|
||||
print '\n'.join(r.keys())
|
||||
|
||||
@click.command()
|
||||
@click.argument('nodes_count')
|
||||
def prepare(nodes_count):
|
||||
r = prepare_nodes(nodes_count)
|
||||
print '\n'.join(r.keys())
|
||||
|
||||
@click.command()
|
||||
@click.argument('node')
|
||||
def add_compute(node):
|
||||
r = create_compute(node)
|
||||
print '\n'.join(r.keys())
|
||||
|
||||
@click.command()
|
||||
@click.argument('node')
|
||||
def add_controller(node):
|
||||
r = create_controller(node)
|
||||
print '\n'.join(r.keys())
|
||||
|
||||
@click.command()
|
||||
def clear():
|
||||
ModelMeta.remove_all()
|
||||
|
||||
|
||||
if __name__ == '__main__':
|
||||
main.add_command(create_all)
|
||||
main.add_command(prepare)
|
||||
main.add_command(add_controller)
|
||||
main.add_command(add_compute)
|
||||
main.add_command(clear)
|
||||
main()
|
||||
|
||||
if PROFILE:
|
||||
pr.disable()
|
||||
s = StringIO.StringIO()
|
||||
sortby = 'cumulative'
|
||||
ps = pstats.Stats(pr, stream=s).sort_stats(sortby)
|
||||
ps.print_stats()
|
||||
print s.getvalue()
|
||||
sys.exit(0)
|
@ -1,22 +0,0 @@
|
||||
id: primary_controller
|
||||
|
||||
resources:
|
||||
- id: rabbit_user
|
||||
from: resources/rabbitmq_user
|
||||
location: {{node}}
|
||||
input:
|
||||
user_name: {{user_name}}
|
||||
password: {{password}}
|
||||
vhost_name: {{vhost_res}}::vhost_name
|
||||
|
||||
updates:
|
||||
- id: {{for}}
|
||||
input:
|
||||
{{for_user}}: rabbit_user::user_name
|
||||
{{for_password}}: rabbit_user::password
|
||||
|
||||
events:
|
||||
- type: depends_on
|
||||
parent_action: rabbit_user.run
|
||||
state: success
|
||||
child_action: {{for}}.update
|
@ -1,79 +0,0 @@
|
||||
#!/usr/bin/env python
|
||||
import requests
|
||||
|
||||
from solar.core.resource import composer as cr
|
||||
from solar.events.api import add_event
|
||||
from solar.events.controls import React
|
||||
|
||||
|
||||
discovery_service = 'http://0.0.0.0:8881'
|
||||
bareon_partitioning = 'http://0.0.0.0:9322/v1/nodes/{0}/partitioning'
|
||||
bareon_repos = 'http://0.0.0.0:9322/v1/nodes/{0}/repos'
|
||||
bareon_sync = 'http://0.0.0.0:9322/v1/actions/sync_all'
|
||||
|
||||
|
||||
class NodeAdapter(dict):
|
||||
|
||||
def __getattr__(self, name):
|
||||
try:
|
||||
return self[name]
|
||||
except KeyError:
|
||||
raise AttributeError(name)
|
||||
|
||||
@property
|
||||
def node_id(self):
|
||||
return self['id']
|
||||
|
||||
@property
|
||||
def partitioning(self):
|
||||
return requests.get(bareon_partitioning.format(self['id'])).json()
|
||||
|
||||
@property
|
||||
def repos(self):
|
||||
return requests.get(bareon_repos.format(self['id'])).json()
|
||||
|
||||
|
||||
# Sync hw info about nodes from discovery service into bareon-api
|
||||
requests.post(bareon_sync)
|
||||
|
||||
# Get list of nodes from discovery service
|
||||
nodes_list = requests.get(discovery_service).json()
|
||||
|
||||
# Create slave node resources
|
||||
node_resources = cr.create('nodes', 'templates/not_provisioned_nodes',
|
||||
{'nodes': nodes_list})
|
||||
|
||||
# Get master node
|
||||
master_node = filter(lambda n: n.name == 'node_master', node_resources)[0]
|
||||
|
||||
with open('/vagrant/tmp/keys/ssh_public') as fp:
|
||||
master_key = fp.read().strip()
|
||||
|
||||
# Dnsmasq resources
|
||||
for node in nodes_list:
|
||||
node = NodeAdapter(node)
|
||||
node_resource = next(n for n in node_resources
|
||||
if n.name.endswith('node_{0}'.format(node.node_id)))
|
||||
|
||||
node_resource.update(
|
||||
{
|
||||
'partitioning': node.partitioning,
|
||||
'master_key': master_key,
|
||||
'repos': node.repos,
|
||||
}
|
||||
)
|
||||
|
||||
dnsmasq = cr.create('dnsmasq_{0}'.format(node.node_id),
|
||||
'resources/dnsmasq', {})[0]
|
||||
master_node.connect(dnsmasq)
|
||||
node_resource.connect(dnsmasq, {'admin_mac': 'exclude_mac_pxe'})
|
||||
|
||||
event = React(node_resource.name, 'run', 'success', node_resource.name,
|
||||
'provision')
|
||||
add_event(event)
|
||||
event = React(node_resource.name, 'provision', 'success', dnsmasq.name,
|
||||
'exclude_mac_pxe')
|
||||
add_event(event)
|
||||
event = React(dnsmasq.name, 'exclude_mac_pxe', 'success',
|
||||
node_resource.name, 'reboot')
|
||||
add_event(event)
|
@ -1,17 +0,0 @@
|
||||
#!/bin/bash
|
||||
|
||||
set -eux
|
||||
|
||||
DIR="$( cd "$( dirname "${BASH_SOURCE[0]}" )" && pwd )"
|
||||
|
||||
# Remove generated pxe exclude files
|
||||
sudo rm -f /etc/dnsmasq.d/no_pxe_*.conf
|
||||
sudo service dnsmasq restart
|
||||
|
||||
solar resource clear_all
|
||||
python "${DIR}"/provision.py
|
||||
|
||||
solar changes stage
|
||||
solar changes process
|
||||
solar orch run-once last
|
||||
watch --color -n1 'solar orch report last'
|
@ -1,47 +0,0 @@
|
||||
Example of 3 node riak cluster.
|
||||
|
||||
At first run:
|
||||
|
||||
`python examples/riak/riaks.py deploy`
|
||||
|
||||
It will prepare riak nodes etc.
|
||||
|
||||
Then you can continue with standard solar things:
|
||||
|
||||
```
|
||||
solar changes stage -d
|
||||
solar changes process
|
||||
solar orch run-once last
|
||||
watch -n 1 solar orch report last
|
||||
```
|
||||
|
||||
Wait until all actions have state `SUCCESS`
|
||||
After that you can add HAProxy on each node:
|
||||
|
||||
`python examples/riak/riaks.py add_haproxies`
|
||||
|
||||
Then again normal solar stuff
|
||||
|
||||
```
|
||||
solar changes stage -d
|
||||
solar changes process
|
||||
solar orch run-once last
|
||||
watch -n 1 solar orch report last
|
||||
```
|
||||
|
||||
|
||||
Wait until all actions have state `SUCCESS`
|
||||
After that you have basic 3 node riak cluster running.
|
||||
|
||||
You can also modify riak http port by:
|
||||
|
||||
`solar resource update riak_service1 riak_port_http=18100`
|
||||
|
||||
And then again standard stuff:
|
||||
|
||||
```
|
||||
solar changes stage -d
|
||||
solar changes process
|
||||
solar orch run-once last
|
||||
watch -n 1 solar orch report last
|
||||
```
|
@ -1,50 +0,0 @@
|
||||
id: haproxy_riak_config
|
||||
|
||||
resources:
|
||||
- id: haproxy_riak_config_http
|
||||
from: resources/haproxy_service_config
|
||||
tags: ['service=riak', 'protocol=http']
|
||||
input:
|
||||
listen_port: #{http_listen_port}#
|
||||
protocol: 'http'
|
||||
name: 'riak_haproxy_http'
|
||||
backends:server:
|
||||
#% for riak in riaks %#
|
||||
- #{riak}#::riak_hostname
|
||||
#% endfor %#
|
||||
backends:port:
|
||||
#% for riak in riaks %#
|
||||
- #{riak}#::riak_port_http
|
||||
#% endfor %#
|
||||
|
||||
- id: haproxy_riak_config_pb
|
||||
from: resources/haproxy_service_config
|
||||
tags: ['service=riak', 'protocol=tcp']
|
||||
input:
|
||||
listen_port: #{pb_listen_port}#
|
||||
protocol: 'tcp'
|
||||
name: 'riak_haproxy_pb'
|
||||
backends:server:
|
||||
#% for riak in riaks %#
|
||||
- #{riak}#::riak_hostname
|
||||
#% endfor %#
|
||||
backends:port:
|
||||
#% for riak in riaks %#
|
||||
- #{riak}#::riak_port_pb
|
||||
#% endfor %#
|
||||
|
||||
updates:
|
||||
- with_tags: ['resource=haproxy_config']
|
||||
input:
|
||||
config:protocol:
|
||||
- haproxy_riak_config_http::protocol
|
||||
- haproxy_riak_config_pb::protocol
|
||||
config:listen_port:
|
||||
- haproxy_riak_config_http::listen_port
|
||||
- haproxy_riak_config_pb::listen_port
|
||||
config:name:
|
||||
- haproxy_riak_config_http::name
|
||||
- haproxy_riak_config_pb::name
|
||||
config:backends:
|
||||
- haproxy_riak_config_http::backends
|
||||
- haproxy_riak_config_pb::backends
|
@ -1,54 +0,0 @@
|
||||
id: riak_cluster
|
||||
|
||||
resources:
|
||||
- id: riak_service1
|
||||
# `./` added by intention
|
||||
from: ./riak_service.yaml
|
||||
input:
|
||||
node: #{nodes[0]}#
|
||||
index: 1
|
||||
join_to: ''
|
||||
|
||||
- id: riak_service2
|
||||
# `./` ommited by intention
|
||||
from: riak_service.yaml
|
||||
input:
|
||||
node: #{nodes[1]}#
|
||||
index: 2
|
||||
join_to: riak_service1
|
||||
|
||||
- id: riak_service3
|
||||
# `./` ommited by intention
|
||||
from: riak_service.yaml
|
||||
input:
|
||||
node: #{nodes[2]}#
|
||||
index: 3
|
||||
join_to: riak_service1
|
||||
|
||||
- id: haproxy_riak_config
|
||||
from: ./haproxy_riak_config.yaml
|
||||
input:
|
||||
http_listen_port: 8098
|
||||
pb_listen_port: 8087
|
||||
riaks: ['riak_service1', 'riak_service2', 'riak_service3']
|
||||
|
||||
- id: haproxy1
|
||||
from: templates/haproxy
|
||||
input:
|
||||
node: #{nodes[0]}#
|
||||
service_configs: ['haproxy_riak_config_pb', 'haproxy_riak_config_http']
|
||||
index: 1
|
||||
|
||||
- id: haproxy2
|
||||
from: templates/haproxy
|
||||
input:
|
||||
node: #{nodes[1]}#
|
||||
service_configs: ['haproxy_riak_config_pb', 'haproxy_riak_config_http']
|
||||
index: 2
|
||||
|
||||
- id: haproxy3
|
||||
from: templates/haproxy
|
||||
input:
|
||||
node: #{nodes[2]}#
|
||||
service_configs: ['haproxy_riak_config_pb', 'haproxy_riak_config_http']
|
||||
index: 3
|
@ -1,61 +0,0 @@
|
||||
id: riak_service
|
||||
|
||||
resources:
|
||||
- id: riak_service#{index}#
|
||||
from: resources/riak_node
|
||||
location: #{node}#
|
||||
input:
|
||||
riak_self_name: riak#{index}#
|
||||
riak_hostname: riak_server#{index}#.solar
|
||||
riak_name: riak#{index}#@riak_server#{index}#.solar
|
||||
#% if join_to %#
|
||||
join_to: #{join_to}#::riak_name
|
||||
#% endif %#
|
||||
ip: #{node}#::ip
|
||||
|
||||
updates:
|
||||
- with_tags: 'resource=hosts_file'
|
||||
input:
|
||||
hosts:name:
|
||||
- riak_service#{index}#::riak_hostname::NO_EVENTS
|
||||
hosts:ip:
|
||||
- riak_service#{index}#::ip::NO_EVENTS
|
||||
|
||||
- with_tags: 'resource=haproxy_service_config & service=riak & protocol=http'
|
||||
input:
|
||||
backends:server:
|
||||
- riak_service#{index}#::riak_hostname
|
||||
backends:port:
|
||||
- riak_service#{index}#::riak_port_http
|
||||
|
||||
- with_tags: 'resource=haproxy_service_config & service=riak & protocol=tcp'
|
||||
input:
|
||||
backends:server:
|
||||
- riak_service#{index}#::riak_hostname
|
||||
backends:port:
|
||||
- riak_service#{index}#::riak_port_pb
|
||||
|
||||
events:
|
||||
- type: depends_on
|
||||
parent:
|
||||
with_tags: 'resource=hosts_file & location=#{node}#'
|
||||
action: run
|
||||
state: success
|
||||
child_action: riak_service#{index}#.run
|
||||
|
||||
#% if join_to %#
|
||||
- type: react_on
|
||||
parent_action: riak_service#{index}#.run
|
||||
state: success
|
||||
child_action: riak_service#{index}#.join
|
||||
|
||||
- type: react_on
|
||||
parent_action: riak_service#{index}#.leave
|
||||
state: success
|
||||
child_action: riak_service#{index}#.join
|
||||
|
||||
- type: react_on
|
||||
parent_action: riak_service#{index}#.join
|
||||
state: success
|
||||
child_action: #{join_to}#.commit
|
||||
#% endif %#
|
@ -1,174 +0,0 @@
|
||||
#!/usr/bin/env python
|
||||
|
||||
# WARNING: this might not be most up-to-date script and not all things might
|
||||
# work here, for most up-to-date version see example-riaks.py
|
||||
# This is just a demo of the template language of Solar
|
||||
|
||||
import click
|
||||
import sys
|
||||
|
||||
from solar.core import resource
|
||||
from solar import template
|
||||
from solar.dblayer.model import ModelMeta
|
||||
|
||||
|
||||
def setup_riak():
|
||||
|
||||
ModelMeta.remove_all()
|
||||
nodes = template.nodes_from('templates/riak_nodes')
|
||||
|
||||
riak_services = nodes.on_each(
|
||||
'resources/riak_node',
|
||||
args={
|
||||
'riak_self_name': 'riak{num}',
|
||||
'riak_hostname': 'riak_server{num}.solar',
|
||||
'riak_name': 'riak{num}@riak_server{num}.solar',
|
||||
}
|
||||
)
|
||||
|
||||
slave_riak_services = riak_services.tail()
|
||||
|
||||
riak_services.take(0).connect_list(
|
||||
slave_riak_services,
|
||||
mapping={
|
||||
'riak_name': 'join_to',
|
||||
}
|
||||
)
|
||||
|
||||
hosts_files = nodes.on_each('resources/hosts_file')
|
||||
|
||||
riak_services.connect_list_to_each(
|
||||
hosts_files,
|
||||
mapping={
|
||||
'ip': 'hosts:ip',
|
||||
'riak_hostname': 'hosts:name',
|
||||
},
|
||||
events=False
|
||||
)
|
||||
|
||||
errors = resource.validate_resources()
|
||||
for r, error in errors:
|
||||
click.echo('ERROR: %s: %s' % (r.name, error))
|
||||
|
||||
if errors:
|
||||
click.echo("ERRORS")
|
||||
sys.exit(1)
|
||||
|
||||
hosts_files.add_deps('run/success', riak_services, 'run')
|
||||
slave_riak_services.add_reacts('run/success', slave_riak_services, 'join')
|
||||
slave_riak_services.add_reacts('leave/success', slave_riak_services, 'join')
|
||||
slave_riak_services.add_react('run/success', riak_services.take(0), 'commit')
|
||||
|
||||
|
||||
def setup_haproxies():
|
||||
# TODO: VR loading needs to be supported, then we can do something like
|
||||
# nodes = template.load('nodes')
|
||||
|
||||
nodes = template.ResourceListTemplate([
|
||||
resource.load('node1'),
|
||||
resource.load('node2'),
|
||||
resource.load('node3'),
|
||||
])
|
||||
riak_services = template.ResourceListTemplate([
|
||||
resource.load('riak_node-0'),
|
||||
resource.load('riak_node-1'),
|
||||
resource.load('riak_node-2'),
|
||||
])
|
||||
|
||||
haproxy_services = nodes.on_each(
|
||||
'resources/haproxy_service'
|
||||
)
|
||||
haproxy_configs = nodes.on_each(
|
||||
'resources/haproxy_config'
|
||||
)
|
||||
haproxy_service_configs_http = riak_services.on_each(
|
||||
'resources/haproxy_service_config',
|
||||
{
|
||||
'listen_port': 8098,
|
||||
'protocol': 'http',
|
||||
'name': 'riak_haproxy_http{num}',
|
||||
}
|
||||
)
|
||||
haproxy_service_configs_pb = riak_services.on_each(
|
||||
'resources/haproxy_service_config',
|
||||
{
|
||||
'listen_port': 8087,
|
||||
'protocol': 'tcp',
|
||||
'name': 'riak_haproxy_pb{num}',
|
||||
}
|
||||
)
|
||||
|
||||
riak_services.connect_list_to_each(
|
||||
haproxy_service_configs_http,
|
||||
{
|
||||
'riak_hostname': 'backends:server',
|
||||
'riak_port_http': 'backends:port',
|
||||
}
|
||||
)
|
||||
riak_services.connect_list_to_each(
|
||||
haproxy_service_configs_pb,
|
||||
{
|
||||
'riak_hostname': 'backends:server',
|
||||
'riak_port_pb': 'backends:port',
|
||||
}
|
||||
)
|
||||
haproxy_service_configs_http.connect_list(
|
||||
haproxy_configs,
|
||||
{
|
||||
'backends': 'config:backends',
|
||||
'listen_port': 'config:listen_port',
|
||||
'protocol': 'config:protocol',
|
||||
'name': 'config:name',
|
||||
}
|
||||
)
|
||||
haproxy_service_configs_pb.connect_list(
|
||||
haproxy_configs,
|
||||
{
|
||||
'backends': 'config:backends',
|
||||
'listen_port': 'config:listen_port',
|
||||
'protocol': 'config:protocol',
|
||||
'name': 'config:name',
|
||||
}
|
||||
)
|
||||
|
||||
#nodes.add_reacts('run/success', haproxy_services, 'install')
|
||||
haproxy_services.add_deps('run/success', haproxy_configs, 'run')
|
||||
haproxy_configs.add_reacts('run/success', haproxy_services, 'apply_config')
|
||||
haproxy_configs.add_reacts('update/success', haproxy_services, 'apply_config')
|
||||
|
||||
errors = resource.validate_resources()
|
||||
for r, error in errors:
|
||||
click.echo('ERROR: %s: %s' % (r.name, error))
|
||||
|
||||
if errors:
|
||||
click.echo("ERRORS")
|
||||
sys.exit(1)
|
||||
|
||||
|
||||
@click.group()
|
||||
def main():
|
||||
pass
|
||||
|
||||
|
||||
@click.command()
|
||||
def deploy():
|
||||
setup_riak()
|
||||
|
||||
|
||||
@click.command()
|
||||
def add_haproxies():
|
||||
setup_haproxies()
|
||||
|
||||
|
||||
@click.command()
|
||||
def undeploy():
|
||||
raise NotImplemented("Not yet")
|
||||
|
||||
|
||||
main.add_command(deploy)
|
||||
main.add_command(undeploy)
|
||||
main.add_command(add_haproxies)
|
||||
|
||||
|
||||
if __name__ == '__main__':
|
||||
main()
|
@ -1,266 +0,0 @@
|
||||
#!/usr/bin/env python
|
||||
|
||||
# To run:
|
||||
# python example-riaks.py deploy
|
||||
# solar changes stage
|
||||
# solar changes process
|
||||
# solar orch run-once last
|
||||
# python example-riaks.py add_haproxies
|
||||
# solar changes stage
|
||||
# solar changes process
|
||||
# solar orch run-once last
|
||||
|
||||
|
||||
import click
|
||||
import sys
|
||||
|
||||
from solar.core import resource
|
||||
from solar.core import signals
|
||||
from solar.core import validation
|
||||
from solar.core.resource import composer as cr
|
||||
from solar import errors
|
||||
|
||||
from solar.dblayer.model import ModelMeta
|
||||
|
||||
from solar.events.controls import React, Dep
|
||||
from solar.events.api import add_event
|
||||
|
||||
from solar.dblayer.solar_models import Resource
|
||||
|
||||
|
||||
def setup_riak():
|
||||
|
||||
ModelMeta.remove_all()
|
||||
resources = cr.create('nodes', 'templates/nodes', {'count': 3})
|
||||
nodes = resources.like('node')
|
||||
hosts_services = resources.like('hosts_file')
|
||||
node1, node2, node3 = nodes
|
||||
|
||||
riak_services = []
|
||||
ips = '10.0.0.%d'
|
||||
for i in xrange(3):
|
||||
num = i + 1
|
||||
r = cr.create('riak_service%d' % num,
|
||||
'resources/riak_node',
|
||||
{'riak_self_name': 'riak%d' % num,
|
||||
'storage_backend': 'leveldb',
|
||||
'riak_hostname': 'riak_server%d.solar' % num})[0]
|
||||
r.connect(r, {'riak_self_name': 'riak_name',
|
||||
'riak_hostname': 'riak_name'})
|
||||
riak_services.append(r)
|
||||
|
||||
for i, riak in enumerate(riak_services):
|
||||
nodes[i].connect(riak)
|
||||
|
||||
for i, riak in enumerate(riak_services[1:]):
|
||||
riak_services[0].connect(riak, {'riak_name': 'join_to'})
|
||||
|
||||
for riak in riak_services:
|
||||
for hosts_file in hosts_services:
|
||||
riak.connect_with_events(hosts_file,
|
||||
{'riak_hostname': 'hosts:name',
|
||||
'ip': 'hosts:ip'})
|
||||
|
||||
Resource.save_all_lazy()
|
||||
errors = resource.validate_resources()
|
||||
for r, error in errors:
|
||||
click.echo('ERROR: %s: %s' % (r.name, error))
|
||||
has_errors = False
|
||||
|
||||
if errors:
|
||||
click.echo("ERRORS")
|
||||
sys.exit(1)
|
||||
|
||||
events = [
|
||||
Dep('hosts_file1', 'run', 'success', 'riak_service1', 'run'),
|
||||
Dep('hosts_file2', 'run', 'success', 'riak_service2', 'run'),
|
||||
Dep('hosts_file3', 'run', 'success', 'riak_service3', 'run'),
|
||||
|
||||
React('riak_service2', 'run', 'success', 'riak_service2', 'join'),
|
||||
React('riak_service3', 'run', 'success', 'riak_service3', 'join'),
|
||||
|
||||
# Dep('riak_service1', 'run', 'success', 'riak_service2', 'join'),
|
||||
# Dep('riak_service1', 'run', 'success', 'riak_service3', 'join'),
|
||||
|
||||
# React('riak_service2', 'join', 'error', 'riak_service2', 'leave'),
|
||||
# React('riak_service3', 'join', 'error', 'riak_service3', 'leave'),
|
||||
|
||||
React('riak_service2', 'leave', 'success', 'riak_service2', 'join'),
|
||||
React('riak_service3', 'leave', 'success', 'riak_service3', 'join'),
|
||||
|
||||
# React('riak_service2', 'leave', 'success', 'riak_service1', 'commit_leave'),
|
||||
# React('riak_service3', 'leave', 'success', 'riak_service1', 'commit_leave'),
|
||||
|
||||
# Dep('riak_service1', 'commit_leave', 'success', 'riak_service2', 'join'),
|
||||
# Dep('riak_service1', 'commit_leave', 'success', 'riak_service3', 'join'),
|
||||
|
||||
React('riak_service3', 'join', 'success', 'riak_service1', 'commit'),
|
||||
React('riak_service2', 'join', 'success', 'riak_service1', 'commit')
|
||||
]
|
||||
|
||||
for event in events:
|
||||
add_event(event)
|
||||
|
||||
click.echo('Use solar changes process & orch')
|
||||
sys.exit(0)
|
||||
|
||||
|
||||
def setup_haproxies():
|
||||
hps = []
|
||||
hpc = []
|
||||
hpsc_http = []
|
||||
hpsc_pb = []
|
||||
for i in xrange(3):
|
||||
num = i + 1
|
||||
hps.append(cr.create('haproxy_service%d' % num,
|
||||
'resources/haproxy_service',
|
||||
{})[0])
|
||||
hpc.append(cr.create('haproxy_config%d' % num,
|
||||
'resources/haproxy_config',
|
||||
{})[0])
|
||||
hpsc_http.append(cr.create('haproxy_service_config_http%d' % num,
|
||||
'resources/haproxy_service_config',
|
||||
{'listen_port': 8098,
|
||||
'protocol': 'http',
|
||||
'name': 'riak_haproxy_http%d' % num})[0])
|
||||
hpsc_pb.append(cr.create('haproxy_service_config_pb%d' % num,
|
||||
'resources/haproxy_service_config',
|
||||
{'listen_port': 8087,
|
||||
'protocol': 'tcp',
|
||||
'name': 'riak_haproxy_pb%d' % num})[0])
|
||||
|
||||
riak1 = resource.load('riak_service1')
|
||||
riak2 = resource.load('riak_service2')
|
||||
riak3 = resource.load('riak_service3')
|
||||
riaks = [riak1, riak2, riak3]
|
||||
|
||||
for single_hpsc in hpsc_http:
|
||||
for riak in riaks:
|
||||
riak.connect(single_hpsc, {
|
||||
'riak_hostname': 'backends:server',
|
||||
'riak_port_http': 'backends:port'})
|
||||
|
||||
for single_hpsc in hpsc_pb:
|
||||
for riak in riaks:
|
||||
riak.connect(single_hpsc,
|
||||
{'riak_hostname': 'backends:server',
|
||||
'riak_port_pb': 'backends:port'})
|
||||
|
||||
# haproxy config to haproxy service
|
||||
|
||||
for single_hpc, single_hpsc in zip(hpc, hpsc_http):
|
||||
single_hpsc.connect(single_hpc, {"backends": "config:backends",
|
||||
"listen_port": "config:listen_port",
|
||||
"protocol": "config:protocol",
|
||||
"name": "config:name"})
|
||||
|
||||
for single_hpc, single_hpsc in zip(hpc, hpsc_pb):
|
||||
single_hpsc.connect(single_hpc, {"backends": "config:backends",
|
||||
"listen_port": "config:listen_port",
|
||||
"protocol": "config:protocol",
|
||||
"name": "config:name"})
|
||||
|
||||
|
||||
# assign haproxy services to each node
|
||||
|
||||
node1 = resource.load('node1')
|
||||
node2 = resource.load('node2')
|
||||
node3 = resource.load('node3')
|
||||
nodes = [node1, node2, node3]
|
||||
|
||||
for single_node, single_hps in zip(nodes, hps):
|
||||
single_node.connect(single_hps)
|
||||
|
||||
for single_node, single_hpc in zip(nodes, hpc):
|
||||
single_node.connect(single_hpc)
|
||||
|
||||
has_errors = False
|
||||
for r in locals().values():
|
||||
|
||||
# TODO: handle list
|
||||
if not isinstance(r, resource.Resource):
|
||||
continue
|
||||
|
||||
# print 'Validating {}'.format(r.name)
|
||||
local_errors = validation.validate_resource(r)
|
||||
if local_errors:
|
||||
has_errors = True
|
||||
print 'ERROR: %s: %s' % (r.name, local_errors)
|
||||
|
||||
if has_errors:
|
||||
print "ERRORS"
|
||||
sys.exit(1)
|
||||
|
||||
events = []
|
||||
for node, single_hps, single_hpc in zip(nodes, hps, hpc):
|
||||
# r = React(node.name, 'run', 'success', single_hps.name, 'install')
|
||||
d = Dep(single_hps.name, 'run', 'success', single_hpc.name, 'run')
|
||||
e1 = React(single_hpc.name, 'run', 'success', single_hps.name, 'apply_config')
|
||||
e2 = React(single_hpc.name, 'update', 'success', single_hps.name, 'apply_config')
|
||||
# events.extend([r, d, e1, e2])
|
||||
events.extend([d, e1, e2])
|
||||
|
||||
for event in events:
|
||||
add_event(event)
|
||||
|
||||
|
||||
@click.command()
|
||||
@click.argument('i', type=int, required=True)
|
||||
def add_solar_agent(i):
|
||||
solar_agent_transport = cr.create('solar_agent_transport%s' % i, 'resources/transport_solar_agent',
|
||||
{'solar_agent_user': 'vagrant',
|
||||
'solar_agent_password': 'password'})[0]
|
||||
transports = resource.load('transports%s' % i)
|
||||
ssh_transport = resource.load('ssh_transport%s' % i)
|
||||
transports_for_solar_agent = cr.create('transports_for_solar_agent%s' % i, 'resources/transports')[0]
|
||||
|
||||
# install solar_agent with ssh
|
||||
signals.connect(transports_for_solar_agent, solar_agent_transport, {})
|
||||
|
||||
signals.connect(ssh_transport, transports_for_solar_agent, {'ssh_key': 'transports:key',
|
||||
'ssh_user': 'transports:user',
|
||||
'ssh_port': 'transports:port',
|
||||
'name': 'transports:name'})
|
||||
|
||||
# add solar_agent to transports on this node
|
||||
signals.connect(solar_agent_transport, transports, {'solar_agent_user': 'transports:user',
|
||||
'solar_agent_port': 'transports:port',
|
||||
'solar_agent_password': 'transports:password',
|
||||
'name': 'transports:name'})
|
||||
|
||||
|
||||
@click.group()
|
||||
def main():
|
||||
pass
|
||||
|
||||
|
||||
@click.command()
|
||||
def deploy():
|
||||
setup_riak()
|
||||
|
||||
|
||||
@click.command()
|
||||
def add_haproxies():
|
||||
setup_haproxies()
|
||||
|
||||
|
||||
@click.command()
|
||||
def undeploy():
|
||||
raise NotImplemented("Not yet")
|
||||
|
||||
|
||||
@click.command()
|
||||
def create_all():
|
||||
setup_riak()
|
||||
setup_haproxies()
|
||||
|
||||
|
||||
main.add_command(deploy)
|
||||
main.add_command(undeploy)
|
||||
main.add_command(add_haproxies)
|
||||
main.add_command(add_solar_agent)
|
||||
main.add_command(create_all)
|
||||
|
||||
|
||||
if __name__ == '__main__':
|
||||
main()
|
@ -1,103 +0,0 @@
|
||||
#!/usr/bin/env python
|
||||
|
||||
# this allows you to create riak cluster as big as you want
|
||||
|
||||
import click
|
||||
import sys
|
||||
|
||||
from solar.core import resource
|
||||
from solar.core import signals
|
||||
from solar.core import validation
|
||||
from solar.core.resource import composer as cr
|
||||
from solar import errors
|
||||
|
||||
from solar.interfaces.db import get_db
|
||||
|
||||
from solar.events.controls import React, Dep
|
||||
from solar.events.api import add_event
|
||||
|
||||
|
||||
db = get_db()
|
||||
|
||||
|
||||
NODES = 3
|
||||
|
||||
def setup_riak(nodes_num=None, hosts_mapping=False):
|
||||
|
||||
if nodes_num is None:
|
||||
nodes_num = NODES
|
||||
db.clear()
|
||||
|
||||
resources = cr.create('nodes', 'templates/nodes', {'count': nodes_num})
|
||||
nodes = [x for x in resources if x.name.startswith('node')]
|
||||
hosts_services = [x for x in resources if x.name.startswith('hosts_file')]
|
||||
|
||||
riak_services = []
|
||||
ips = '10.0.0.%d'
|
||||
for i in xrange(nodes_num):
|
||||
num = i + 1
|
||||
r = cr.create('riak_service%d' % num,
|
||||
'resources/riak_node',
|
||||
{'riak_self_name': 'riak%d' % num,
|
||||
'riak_hostname': 'riak_server%d.solar' % num,
|
||||
'riak_name': 'riak%d@riak_server%d.solar' % (num, num)})[0]
|
||||
riak_services.append(r)
|
||||
|
||||
for i, riak in enumerate(riak_services):
|
||||
nodes[i].connect(riak)
|
||||
|
||||
for i, riak in enumerate(riak_services[1:]):
|
||||
riak_services[0].connect(riak, {'riak_name': 'join_to'})
|
||||
|
||||
if hosts_mapping:
|
||||
for riak in riak_services:
|
||||
for hosts_file in hosts_services:
|
||||
riak.connect_with_events(hosts_file,
|
||||
{'riak_hostname': 'hosts:name',
|
||||
'ip': 'hosts:ip'})
|
||||
|
||||
res_errors = resource.validate_resources()
|
||||
for r, error in res_errors:
|
||||
click.echo('ERROR: %s: %s' % (r.name, error))
|
||||
has_errors = False
|
||||
|
||||
if has_errors:
|
||||
click.echo("ERRORS")
|
||||
sys.exit(1)
|
||||
|
||||
events = []
|
||||
for x in xrange(nodes_num):
|
||||
i = x + 1
|
||||
if hosts_mapping:
|
||||
events.append(Dep('hosts_file%d' % i, 'run', 'success', 'riak_service%d' % i, 'run'))
|
||||
if i >= 2:
|
||||
events.append(React('riak_service%d' % i, 'run', 'success', 'riak_service%d' % i, 'join'))
|
||||
events.append(React('riak_service%d' % i, 'join', 'success', 'riak_service1', 'commit'))
|
||||
|
||||
for event in events:
|
||||
add_event(event)
|
||||
|
||||
click.echo('Use solar changes process & orch')
|
||||
sys.exit(0)
|
||||
|
||||
|
||||
@click.group()
|
||||
def main():
|
||||
pass
|
||||
|
||||
|
||||
@click.command()
|
||||
@click.argument('nodes_count', type=int)
|
||||
@click.argument('hosts_mapping', type=bool)
|
||||
def deploy(nodes_count, hosts_mapping):
|
||||
click.secho("With big nodes_count, this example is DB heavy, it creates NxN connections, continue ? [y/N] ", fg='red', nl=False)
|
||||
c= click.getchar()
|
||||
if c in ('y', 'Y'):
|
||||
setup_riak(nodes_count, hosts_mapping)
|
||||
else:
|
||||
click.echo("Aborted")
|
||||
|
||||
|
||||
if __name__ == '__main__':
|
||||
main.add_command(deploy)
|
||||
main()
|
@ -1,61 +0,0 @@
|
||||
import click
|
||||
import sys
|
||||
import time
|
||||
|
||||
from solar.core import resource
|
||||
from solar.core import signals
|
||||
from solar.core.resource import composer as cr
|
||||
from solar.dblayer.model import ModelMeta
|
||||
|
||||
|
||||
def run():
|
||||
ModelMeta.remove_all()
|
||||
|
||||
node = cr.create('node', 'resources/ro_node', {'name': 'first' + str(time.time()),
|
||||
'ip': '10.0.0.3',
|
||||
'node_id': 'node1',
|
||||
})[0]
|
||||
|
||||
transports = cr.create('transports_node1', 'resources/transports')[0]
|
||||
transports_for_solar_agent = cr.create('transports_for_solar_agent', 'resources/transports')[0]
|
||||
|
||||
ssh_transport = cr.create('ssh_transport', 'resources/transport_ssh',
|
||||
{'ssh_key': '/vagrant/.vagrant/machines/solar-dev1/virtualbox/private_key',
|
||||
'ssh_user': 'vagrant'})[0]
|
||||
|
||||
solar_agent_transport = cr.create('solar_agent_transport', 'resources/transport_solar_agent',
|
||||
{'solar_agent_user': 'vagrant',
|
||||
'solar_agent_password': 'password'})[0]
|
||||
|
||||
transports_for_solar_agent.connect(solar_agent_transport, {})
|
||||
ssh_transport.connect(transports_for_solar_agent,{'ssh_key': 'transports:key',
|
||||
'ssh_user': 'transports:user',
|
||||
'ssh_port': 'transports:port',
|
||||
'name': 'transports:name'})
|
||||
# set transports_id
|
||||
transports.connect(node, {})
|
||||
|
||||
# it uses reverse mappings
|
||||
ssh_transport.connect(transports, {'ssh_key': 'transports:key',
|
||||
'ssh_user': 'transports:user',
|
||||
'ssh_port': 'transports:port',
|
||||
'name': 'transports:name'})
|
||||
solar_agent_transport.connect(transports, {'solar_agent_user': 'transports:user',
|
||||
'solar_agent_port': 'transports:port',
|
||||
'solar_agent_password': 'transports:password',
|
||||
'name': 'transports:name'})
|
||||
|
||||
|
||||
hosts = cr.create('hosts_file', 'resources/hosts_file', {})[0]
|
||||
node.connect(hosts, {
|
||||
'ip': 'hosts:ip',
|
||||
'name': 'hosts:name'
|
||||
})
|
||||
|
||||
# for r in (node, hosts, ssh_transport, transports):
|
||||
# print r.name, repr(r.args['location_id']), repr(r.args['transports_id'])
|
||||
|
||||
# print hosts.transports()
|
||||
# print hosts.ip()
|
||||
|
||||
run()
|
@ -1,25 +0,0 @@
|
||||
Example of using torrent transport with solar. Torrent is used to distribute task data. After fetching is finished torrent client forks and continues seeding.
|
||||
|
||||
|
||||
The example contains single node with single host mapping + transports.
|
||||
|
||||
Execute:
|
||||
```
|
||||
python examples/torrent/example.py
|
||||
solar changes stage
|
||||
solar changes process
|
||||
solar orch run-once last
|
||||
```
|
||||
|
||||
Wait for finish:
|
||||
|
||||
```
|
||||
solar orch report last -w 100
|
||||
```
|
||||
|
||||
After this you should see new entry in `/etc/hosts` file.
|
||||
|
||||
|
||||
* All created torrents are in `/vagrant/torrents`, it doesn't need to be shared
|
||||
* Initial seeding is done using torrent file
|
||||
* Downloading and then seeding is always done with magnetlinks
|
@ -1,70 +0,0 @@
|
||||
import time
|
||||
|
||||
from solar.core.resource import composer as cr
|
||||
from solar import errors
|
||||
from solar.dblayer.model import ModelMeta
|
||||
|
||||
|
||||
def run():
|
||||
ModelMeta.remove_all()
|
||||
|
||||
node = cr.create('node', 'resources/ro_node', {'name': 'first' + str(time.time()),
|
||||
'ip': '10.0.0.3',
|
||||
'node_id': 'node1',
|
||||
})[0]
|
||||
|
||||
transports = cr.create('transports_node1', 'resources/transports')[0]
|
||||
|
||||
ssh_transport = cr.create('ssh_transport', 'resources/transport_ssh',
|
||||
{'ssh_key': '/vagrant/.vagrant/machines/solar-dev1/virtualbox/private_key',
|
||||
'ssh_user': 'vagrant'})[0]
|
||||
|
||||
transports.connect(node, {})
|
||||
|
||||
# it uses reverse mappings
|
||||
ssh_transport.connect(transports, {'ssh_key': 'transports:key',
|
||||
'ssh_user': 'transports:user',
|
||||
'ssh_port': 'transports:port',
|
||||
'name': 'transports:name'})
|
||||
|
||||
hosts = cr.create('hosts_file', 'resources/hosts_file', {})[0]
|
||||
|
||||
# let's add torrent transport for hosts file deployment (useless in real life)
|
||||
|
||||
torrent_transport = cr.create('torrent_transport',
|
||||
'resources/transport_torrent',
|
||||
{'trackers': ['udp://open.demonii.com:1337',
|
||||
'udp://tracker.openbittorrent.com:80']})[0]
|
||||
# you could use any trackers as you want
|
||||
|
||||
transports_for_torrent = cr.create(
|
||||
'transports_for_torrent', 'resources/transports')[0]
|
||||
|
||||
transports_for_torrent.connect(torrent_transport, {})
|
||||
|
||||
ssh_transport.connect_with_events(transports_for_torrent, {'ssh_key': 'transports:key',
|
||||
'ssh_user': 'transports:user',
|
||||
'ssh_port': 'transports:port',
|
||||
'name': 'transports:name'},
|
||||
events={})
|
||||
|
||||
transports_for_hosts = cr.create(
|
||||
'transports_for_hosts', 'resources/transports')[0]
|
||||
|
||||
torrent_transport.connect(transports_for_hosts, {'trackers': 'transports:trackers',
|
||||
'name': 'transports:name'})
|
||||
|
||||
ssh_transport.connect(transports_for_hosts, {'ssh_key': 'transports:key',
|
||||
'ssh_user': 'transports:user',
|
||||
'ssh_port': 'transports:port',
|
||||
'name': 'transports:name'})
|
||||
|
||||
transports_for_hosts.connect(hosts)
|
||||
transports_for_hosts.connect_with_events(node, events={})
|
||||
|
||||
node.connect(hosts, {
|
||||
'ip': 'hosts:ip',
|
||||
'name': 'hosts:name'
|
||||
})
|
||||
|
||||
run()
|
@ -1,9 +0,0 @@
|
||||
- hosts: localhost
|
||||
sudo: yes
|
||||
vars:
|
||||
var1: 'playbook'
|
||||
roles:
|
||||
- { role: "test_role" }
|
||||
tasks:
|
||||
- debug: msg="VAR1 value is {{var1}}"
|
||||
- fail: msg='just test failure'
|
@ -1,4 +0,0 @@
|
||||
|
||||
var1: initial
|
||||
uuid: stuff
|
||||
def1: the_same
|
@ -1 +0,0 @@
|
||||
- debug: msg="Variable1 {{ var1 }} with uuid {{ uuid }} and default var {{ def1 }}"
|
@ -1,10 +0,0 @@
|
||||
handler: ansible_playbook
|
||||
version: 0.0.1
|
||||
input:
|
||||
var1:
|
||||
type: str!
|
||||
value: meta
|
||||
uuid:
|
||||
type: str!
|
||||
value: 'aa1das1231'
|
||||
|
@ -1,6 +0,0 @@
|
||||
- hosts: '*'
|
||||
sudo: yes
|
||||
vars:
|
||||
default1: playbook
|
||||
tasks:
|
||||
- debug: msg="my message {{default1}}"
|
@ -1,15 +0,0 @@
|
||||
handler: ansible_playbook
|
||||
version: 0.0.1
|
||||
input:
|
||||
ip:
|
||||
type: str!
|
||||
value:
|
||||
# ssh_user:
|
||||
# type: str!
|
||||
# value:
|
||||
# ssh_key:
|
||||
# type: str!
|
||||
# value:
|
||||
default1:
|
||||
type: str!
|
||||
value: meta
|
@ -1,4 +0,0 @@
|
||||
# Apache puppet resource
|
||||
|
||||
This class installs Apache and manages apache service.
|
||||
Defaults provided for Debian OS family.
|
@ -1,5 +0,0 @@
|
||||
class {'apache':
|
||||
service_enable => false,
|
||||
service_ensure => 'stopped',
|
||||
package_ensure => 'absent',
|
||||
}
|
@ -1,120 +0,0 @@
|
||||
$resource = hiera($::resource_name)
|
||||
|
||||
$apache_name = $resource['input']['apache_name']
|
||||
$service_name = $resource['input']['service_name']
|
||||
$default_mods = $resource['input']['default_mods']
|
||||
$default_vhost = $resource['input']['default_vhost']
|
||||
$default_charset = $resource['input']['default_charset']
|
||||
$default_confd_files = $resource['input']['default_confd_files']
|
||||
$default_ssl_vhost = $resource['input']['default_ssl_vhost']
|
||||
$default_ssl_cert = $resource['input']['default_ssl_cert']
|
||||
$default_ssl_key = $resource['input']['default_ssl_key']
|
||||
$default_ssl_chain = $resource['input']['default_ssl_chain']
|
||||
$default_ssl_ca = $resource['input']['default_ssl_ca']
|
||||
$default_ssl_crl_path = $resource['input']['default_ssl_crl_path']
|
||||
$default_ssl_crl = $resource['input']['default_ssl_crl']
|
||||
$default_ssl_crl_check = $resource['input']['default_ssl_crl_check']
|
||||
$default_type = $resource['input']['default_type']
|
||||
$ip = $resource['input']['ip']
|
||||
$service_restart = $resource['input']['service_restart']
|
||||
$purge_configs = $resource['input']['purge_configs']
|
||||
$purge_vhost_dir = $resource['input']['purge_vhost_dir']
|
||||
$purge_vdir = $resource['input']['purge_vdir']
|
||||
$serveradmin = $resource['input']['serveradmin']
|
||||
$sendfile = $resource['input']['sendfile']
|
||||
$error_documents = $resource['input']['error_documents']
|
||||
$timeout = $resource['input']['timeout']
|
||||
$httpd_dir = $resource['input']['httpd_dir']
|
||||
$server_root = $resource['input']['server_root']
|
||||
$conf_dir = $resource['input']['conf_dir']
|
||||
$confd_dir = $resource['input']['confd_dir']
|
||||
$vhost_dir = $resource['input']['vhost_dir']
|
||||
$vhost_enable_dir = $resource['input']['vhost_enable_dir']
|
||||
$mod_dir = $resource['input']['mod_dir']
|
||||
$mod_enable_dir = $resource['input']['mod_enable_dir']
|
||||
$mpm_module = $resource['input']['mpm_module']
|
||||
$lib_path = $resource['input']['lib_path']
|
||||
$conf_template = $resource['input']['conf_template']
|
||||
$servername = $resource['input']['servername']
|
||||
$manage_user = $resource['input']['manage_user']
|
||||
$manage_group = $resource['input']['manage_group']
|
||||
$user = $resource['input']['user']
|
||||
$group = $resource['input']['group']
|
||||
$keepalive = $resource['input']['keepalive']
|
||||
$keepalive_timeout = $resource['input']['keepalive_timeout']
|
||||
$max_keepalive_requests = $resource['input']['max_keepalive_requests']
|
||||
$logroot = $resource['input']['logroot']
|
||||
$logroot_mode = $resource['input']['logroot_mode']
|
||||
$log_level = $resource['input']['log_level']
|
||||
$log_formats = $resource['input']['log_formats']
|
||||
$ports_file = $resource['input']['ports_file']
|
||||
$docroot = $resource['input']['docroot']
|
||||
$apache_version = $resource['input']['apache_version']
|
||||
$server_tokens = $resource['input']['server_tokens']
|
||||
$server_signature = $resource['input']['server_signature']
|
||||
$trace_enable = $resource['input']['trace_enable']
|
||||
$allow_encoded_slashes = $resource['input']['allow_encoded_slashes']
|
||||
$package_ensure = $resource['input']['package_ensure']
|
||||
$use_optional_includes = $resource['input']['use_optional_includes']
|
||||
|
||||
class {'apache':
|
||||
apache_name => $apache_name,
|
||||
service_name => $service_name,
|
||||
default_mods => $default_mods,
|
||||
default_vhost => $default_vhost,
|
||||
default_charset => $default_charset,
|
||||
default_confd_files => $default_confd_files,
|
||||
default_ssl_vhost => $default_ssl_vhost,
|
||||
default_ssl_cert => $default_ssl_cert,
|
||||
default_ssl_key => $default_ssl_key,
|
||||
default_ssl_chain => $default_ssl_chain,
|
||||
default_ssl_ca => $default_ssl_ca,
|
||||
default_ssl_crl_path => $default_ssl_crl_path,
|
||||
default_ssl_crl => $default_ssl_crl,
|
||||
default_ssl_crl_check => $default_ssl_crl_check,
|
||||
default_type => $default_type,
|
||||
ip => $ip,
|
||||
service_enable => true,
|
||||
service_manage => true,
|
||||
service_ensure => 'running',
|
||||
service_restart => $service_restart,
|
||||
purge_configs => $purge_configs,
|
||||
purge_vhost_dir => $purge_vhost_dir,
|
||||
purge_vdir => $purge_vdir,
|
||||
serveradmin => $serveradmin,
|
||||
sendfile => $sendfile,
|
||||
error_documents => $error_documents,
|
||||
timeout => $timeout,
|
||||
httpd_dir => $httpd_dir,
|
||||
server_root => $server_root,
|
||||
conf_dir => $conf_dir,
|
||||
confd_dir => $confd_dir,
|
||||
vhost_dir => $vhost_dir,
|
||||
vhost_enable_dir => $vhost_enable_dir,
|
||||
mod_dir => $mod_dir,
|
||||
mod_enable_dir => $mod_enable_dir,
|
||||
mpm_module => $mpm_module,
|
||||
lib_path => $lib_path,
|
||||
conf_template => $conf_template,
|
||||
servername => $servername,
|
||||
manage_user => $manage_user,
|
||||
manage_group => $manage_group,
|
||||
user => $user,
|
||||
group => $group,
|
||||
keepalive => $keepalive,
|
||||
keepalive_timeout => $keepalive_timeout,
|
||||
max_keepalive_requests => $max_keepalive_requests,
|
||||
logroot => $logroot,
|
||||
logroot_mode => $logroot_mode,
|
||||
log_level => $log_level,
|
||||
log_formats => $log_formats,
|
||||
ports_file => $ports_file,
|
||||
docroot => $docroot,
|
||||
apache_version => $apache_version,
|
||||
server_tokens => $server_tokens,
|
||||
server_signature => $server_signature,
|
||||
trace_enable => $trace_enable,
|
||||
allow_encoded_slashes => $allow_encoded_slashes,
|
||||
package_ensure => $package_ensure,
|
||||
use_optional_includes => $use_optional_includes,
|
||||
}
|
@ -1,184 +0,0 @@
|
||||
handler: puppet
|
||||
version: 1.0.0
|
||||
input:
|
||||
apache_name:
|
||||
schema: str
|
||||
value: 'apache2'
|
||||
service_name:
|
||||
schema: str
|
||||
value: 'apache2'
|
||||
default_mods:
|
||||
schema: bool
|
||||
value: true
|
||||
default_vhost:
|
||||
schema: bool
|
||||
value: true
|
||||
default_charset:
|
||||
schema: str
|
||||
value:
|
||||
default_confd_files:
|
||||
schema: bool
|
||||
value: true
|
||||
default_ssl_vhost:
|
||||
schema: bool
|
||||
value: false
|
||||
default_ssl_cert:
|
||||
schema: str
|
||||
value: '/etc/ssl/certs/ssl-cert-snakeoil.pem'
|
||||
default_ssl_key:
|
||||
schema: str
|
||||
value: '/etc/ssl/private/ssl-cert-snakeoil.key'
|
||||
default_ssl_chain:
|
||||
schema: str
|
||||
value:
|
||||
default_ssl_ca:
|
||||
schema: str
|
||||
value:
|
||||
default_ssl_crl_path:
|
||||
schema: str
|
||||
value:
|
||||
default_ssl_crl:
|
||||
schema: str
|
||||
value:
|
||||
default_ssl_crl_check:
|
||||
schema: str
|
||||
value:
|
||||
default_type:
|
||||
schema: str
|
||||
value: 'none'
|
||||
service_restart:
|
||||
schema: str
|
||||
value: 'restart'
|
||||
purge_configs:
|
||||
schema: bool
|
||||
value: true
|
||||
purge_vhost_dir:
|
||||
schema: str
|
||||
value:
|
||||
purge_vdir:
|
||||
schema: bool
|
||||
value: false
|
||||
serveradmin:
|
||||
schema: str
|
||||
value: 'root@localhost'
|
||||
sendfile:
|
||||
schema: str
|
||||
value: 'On'
|
||||
error_documents:
|
||||
schema: bool
|
||||
value: false
|
||||
timeout:
|
||||
schema: int
|
||||
value: 120
|
||||
httpd_dir:
|
||||
schema: str
|
||||
value: '/etc/apache2'
|
||||
server_root:
|
||||
schema: str
|
||||
value: '/etc/apache2'
|
||||
conf_dir:
|
||||
schema: str
|
||||
value: '/etc/apache2'
|
||||
confd_dir:
|
||||
schema: str
|
||||
value: '/etc/apache2/conf.d'
|
||||
vhost_dir:
|
||||
schema: str
|
||||
value: '/etc/apache2/sites-available'
|
||||
vhost_enable_dir:
|
||||
schema: str
|
||||
value: '/etc/apache2/sites-enabled'
|
||||
mod_dir:
|
||||
schema: str
|
||||
value: '/etc/apache2/mods-available'
|
||||
mod_enable_dir:
|
||||
schema: str
|
||||
value: '/etc/apache2/mods-enabled'
|
||||
mpm_module:
|
||||
schema: str
|
||||
value: 'worker'
|
||||
lib_path:
|
||||
schema: str
|
||||
value: '/usr/lib/apache2/modules'
|
||||
conf_template:
|
||||
schema: str
|
||||
value: 'apache/httpd.conf.erb'
|
||||
servername:
|
||||
schema: str!
|
||||
value:
|
||||
manage_user:
|
||||
schema: bool
|
||||
value: true
|
||||
manage_group:
|
||||
schema: bool
|
||||
value: true
|
||||
user:
|
||||
schema: str
|
||||
value: 'www-data'
|
||||
group:
|
||||
schema: str
|
||||
value: 'www-data'
|
||||
keepalive:
|
||||
schema: str
|
||||
value: 'Off'
|
||||
keepalive_timeout:
|
||||
schema: int
|
||||
value: 15
|
||||
max_keepalive_requests:
|
||||
schema: int
|
||||
value: 100
|
||||
logroot:
|
||||
schema: str
|
||||
value: '/var/log/apache2'
|
||||
logroot_mode:
|
||||
schema: str
|
||||
value: '0640'
|
||||
log_level:
|
||||
schema: str
|
||||
value: 'warn'
|
||||
log_formats:
|
||||
schema: {}
|
||||
value: {}
|
||||
ports_file:
|
||||
schema: str
|
||||
value: '/etc/apache2/ports.conf'
|
||||
docroot:
|
||||
schema: str
|
||||
value: '/srv/www'
|
||||
apache_version:
|
||||
schema: str
|
||||
value: '2.4'
|
||||
server_tokens:
|
||||
schema: str
|
||||
value: 'OS'
|
||||
server_signature:
|
||||
schema: str
|
||||
value: 'On'
|
||||
trace_enable:
|
||||
schema: str
|
||||
value: 'On'
|
||||
allow_encoded_slashes:
|
||||
schema: str
|
||||
value:
|
||||
package_ensure:
|
||||
schema: str
|
||||
value: 'installed'
|
||||
use_optional_includes:
|
||||
schema: bool
|
||||
value: false
|
||||
|
||||
git:
|
||||
schema: {repository: str!, branch: str!}
|
||||
value: {repository: 'https://github.com/puppetlabs/puppetlabs-apache.git', branch: '1.5.0'}
|
||||
|
||||
ip:
|
||||
schema: str!
|
||||
value:
|
||||
# ssh_key:
|
||||
# schema: str!
|
||||
# value:
|
||||
# ssh_user:
|
||||
# schema: str!
|
||||
# value:
|
||||
|
||||
tags: [resource/apache_service, resources/apache]
|
@ -1,11 +0,0 @@
|
||||
import requests
|
||||
|
||||
from solar.core.log import log
|
||||
|
||||
|
||||
def test(resource):
|
||||
log.debug('Testing apache_puppet')
|
||||
requests.get(
|
||||
'http://%s:%s' % (resource.args['ip'], 80)
|
||||
|
||||
)
|
@ -1,9 +0,0 @@
|
||||
- hosts: [{{host}}]
|
||||
sudo: yes
|
||||
tasks:
|
||||
- shell: rm -f {{item}}
|
||||
with_items:
|
||||
- /etc/apt/sources.list.d/{{name}}.list
|
||||
- /etc/apt/preferences.d/{{name}}.pref
|
||||
- shell: apt-get update
|
||||
when: {{validate_integrity}}
|
@ -1,11 +0,0 @@
|
||||
- hosts: [{{host}}]
|
||||
sudo: yes
|
||||
tasks:
|
||||
- template:
|
||||
src: {{templates_dir}}/source
|
||||
dest: /etc/apt/sources.list.d/{{name}}.list
|
||||
- template:
|
||||
src: {{templates_dir}}/preferences
|
||||
dest: /etc/apt/preferences.d/{{name}}.pref
|
||||
- shell: apt-get update
|
||||
when: {{validate_integrity}}
|
@ -1,24 +0,0 @@
|
||||
handler: ansible
|
||||
version: 1.0.0
|
||||
input:
|
||||
ip:
|
||||
schema: str!
|
||||
value:
|
||||
repo:
|
||||
schema: str!
|
||||
value:
|
||||
name:
|
||||
schema: str!
|
||||
value:
|
||||
package:
|
||||
schema: str
|
||||
value: '*'
|
||||
pin:
|
||||
schema: str
|
||||
value:
|
||||
pin_priority:
|
||||
schema: int
|
||||
value:
|
||||
validate_integrity:
|
||||
schema: bool
|
||||
value: true
|
@ -1,3 +0,0 @@
|
||||
Package: {{package}}
|
||||
Pin: {{pin}}
|
||||
Pin-Priority: {{pin_priority}}
|
@ -1 +0,0 @@
|
||||
{{repo}}
|
@ -1,17 +0,0 @@
|
||||
#!/bin/sh
|
||||
|
||||
BASE_PATH={{ target_directory }}
|
||||
KEY_NAME={{ key_name }}
|
||||
|
||||
function generate_ssh_keys {
|
||||
local dir_path=$BASE_PATH$KEY_NAME/
|
||||
local key_path=$dir_path$KEY_NAME
|
||||
mkdir -p $dir_path
|
||||
if [ ! -f $key_path ]; then
|
||||
ssh-keygen -b 2048 -t rsa -N '' -f $key_path 2>&1
|
||||
else
|
||||
echo 'Key $key_path already exists'
|
||||
fi
|
||||
}
|
||||
|
||||
generate_ssh_keys
|
@ -1,16 +0,0 @@
|
||||
handler: shell
|
||||
version: 1.0.0
|
||||
input:
|
||||
ip:
|
||||
schema: str!
|
||||
value:
|
||||
target_directory:
|
||||
schema: str!
|
||||
value: /var/lib/astute/
|
||||
key_name:
|
||||
schema: str!
|
||||
value: ceph
|
||||
path:
|
||||
schema: str!
|
||||
value: /var/lib/astute/ceph/
|
||||
tags: []
|
@ -1,95 +0,0 @@
|
||||
notice('MODULAR: ceph/mon.pp')
|
||||
|
||||
|
||||
$storage_hash = hiera('storage', {})
|
||||
$public_vip = hiera('public_vip')
|
||||
$management_vip = hiera('management_vip')
|
||||
$use_syslog = hiera('use_syslog', true)
|
||||
$syslog_log_facility_ceph = hiera('syslog_log_facility_ceph','LOG_LOCAL0')
|
||||
$keystone_hash = hiera('keystone', {})
|
||||
$mon_address_map = get_node_to_ipaddr_map_by_network_role(hiera_hash('ceph_monitor_nodes'), 'ceph/public')
|
||||
|
||||
if ($storage_hash['images_ceph']) {
|
||||
$glance_backend = 'ceph'
|
||||
} elsif ($storage_hash['images_vcenter']) {
|
||||
$glance_backend = 'vmware'
|
||||
} else {
|
||||
$glance_backend = 'swift'
|
||||
}
|
||||
|
||||
if ($storage_hash['volumes_ceph'] or
|
||||
$storage_hash['images_ceph'] or
|
||||
$storage_hash['objects_ceph'] or
|
||||
$storage_hash['ephemeral_ceph']
|
||||
) {
|
||||
$use_ceph = true
|
||||
} else {
|
||||
$use_ceph = false
|
||||
}
|
||||
|
||||
if $use_ceph {
|
||||
$ceph_primary_monitor_node = hiera('ceph_primary_monitor_node')
|
||||
$primary_mons = keys($ceph_primary_monitor_node)
|
||||
$primary_mon = $ceph_primary_monitor_node[$primary_mons[0]]['name']
|
||||
|
||||
prepare_network_config(hiera_hash('network_scheme'))
|
||||
$ceph_cluster_network = get_network_role_property('ceph/replication', 'network')
|
||||
$ceph_public_network = get_network_role_property('ceph/public', 'network')
|
||||
$mon_addr = get_network_role_property('ceph/public', 'ipaddr')
|
||||
|
||||
class {'ceph':
|
||||
primary_mon => $primary_mon,
|
||||
mon_hosts => keys($mon_address_map),
|
||||
mon_ip_addresses => values($mon_address_map),
|
||||
mon_addr => $mon_addr,
|
||||
cluster_node_address => $public_vip,
|
||||
osd_pool_default_size => $storage_hash['osd_pool_size'],
|
||||
osd_pool_default_pg_num => $storage_hash['pg_num'],
|
||||
osd_pool_default_pgp_num => $storage_hash['pg_num'],
|
||||
use_rgw => false,
|
||||
glance_backend => $glance_backend,
|
||||
rgw_pub_ip => $public_vip,
|
||||
rgw_adm_ip => $management_vip,
|
||||
rgw_int_ip => $management_vip,
|
||||
cluster_network => $ceph_cluster_network,
|
||||
public_network => $ceph_public_network,
|
||||
use_syslog => $use_syslog,
|
||||
syslog_log_level => hiera('syslog_log_level_ceph', 'info'),
|
||||
syslog_log_facility => $syslog_log_facility_ceph,
|
||||
rgw_keystone_admin_token => $keystone_hash['admin_token'],
|
||||
ephemeral_ceph => $storage_hash['ephemeral_ceph']
|
||||
}
|
||||
|
||||
if ($storage_hash['volumes_ceph']) {
|
||||
include ::cinder::params
|
||||
service { 'cinder-volume':
|
||||
ensure => 'running',
|
||||
name => $::cinder::params::volume_service,
|
||||
hasstatus => true,
|
||||
hasrestart => true,
|
||||
}
|
||||
|
||||
service { 'cinder-backup':
|
||||
ensure => 'running',
|
||||
name => $::cinder::params::backup_service,
|
||||
hasstatus => true,
|
||||
hasrestart => true,
|
||||
}
|
||||
|
||||
Class['ceph'] ~> Service['cinder-volume']
|
||||
Class['ceph'] ~> Service['cinder-backup']
|
||||
}
|
||||
|
||||
if ($storage_hash['images_ceph']) {
|
||||
include ::glance::params
|
||||
service { 'glance-api':
|
||||
ensure => 'running',
|
||||
name => $::glance::params::api_service_name,
|
||||
hasstatus => true,
|
||||
hasrestart => true,
|
||||
}
|
||||
|
||||
Class['ceph'] ~> Service['glance-api']
|
||||
}
|
||||
|
||||
}
|
@ -1,4 +0,0 @@
|
||||
prepare_network_config(hiera_hash('network_scheme'))
|
||||
$ceph_cluster_network = get_network_role_property('ceph/replication', 'network')
|
||||
|
||||
notify{"The value is: ${ceph_cluster_network}": }
|
@ -1,37 +0,0 @@
|
||||
handler: puppetv2
|
||||
version: 1.0.0
|
||||
input:
|
||||
ip:
|
||||
schema: str!
|
||||
value:
|
||||
public_vip:
|
||||
schema: str!
|
||||
value:
|
||||
management_vip:
|
||||
schema: str!
|
||||
value:
|
||||
use_syslog:
|
||||
schema: bool
|
||||
value: true
|
||||
keystone:
|
||||
schema: {'admin_token': 'str'}
|
||||
value: {}
|
||||
ceph_monitor_nodes:
|
||||
schema: []
|
||||
value: []
|
||||
ceph_primary_monitor_node:
|
||||
schema: []
|
||||
value: []
|
||||
storage:
|
||||
schema: {}
|
||||
value: {}
|
||||
network_scheme:
|
||||
schema: {}
|
||||
value: {}
|
||||
role:
|
||||
schema: str!
|
||||
value:
|
||||
puppet_modules:
|
||||
schema: str!
|
||||
value:
|
||||
tags: []
|
@ -1,98 +0,0 @@
|
||||
# Cinder API resource for puppet handler
|
||||
|
||||
Setup and configure the cinder API endpoint
|
||||
|
||||
## Parameters
|
||||
|
||||
source https://github.com/openstack/puppet-cinder/blob/5.1.0/manifests/api.pp
|
||||
|
||||
``keystone_password``
|
||||
The password to use for authentication (keystone)
|
||||
|
||||
``keystone_enabled``
|
||||
(optional) Use keystone for authentification
|
||||
Defaults to true
|
||||
|
||||
``keystone_tenant``
|
||||
(optional) The tenant of the auth user
|
||||
Defaults to services
|
||||
|
||||
``keystone_user``
|
||||
(optional) The name of the auth user
|
||||
Defaults to cinder
|
||||
|
||||
``keystone_auth_host``
|
||||
(optional) The keystone host
|
||||
Defaults to localhost
|
||||
|
||||
``keystone_auth_port``
|
||||
(optional) The keystone auth port
|
||||
Defaults to 35357
|
||||
|
||||
``keystone_auth_protocol``
|
||||
(optional) The protocol used to access the auth host
|
||||
Defaults to http.
|
||||
|
||||
``os_region_name``
|
||||
(optional) Some operations require cinder to make API requests
|
||||
to Nova. This sets the keystone region to be used for these
|
||||
requests. For example, boot-from-volume.
|
||||
Defaults to undef.
|
||||
|
||||
``keystone_auth_admin_prefix``
|
||||
(optional) The admin_prefix used to admin endpoint of the auth host
|
||||
This allow admin auth URIs like http://auth_host:35357/keystone.
|
||||
(where '/keystone' is the admin prefix)
|
||||
Defaults to false for empty. If defined, should be a string with a
|
||||
leading '/' and no trailing '/'.
|
||||
|
||||
``service_port``
|
||||
(optional) The cinder api port
|
||||
Defaults to 5000
|
||||
|
||||
``service_workers``
|
||||
(optional) Number of cinder-api workers
|
||||
Defaults to $::processorcount
|
||||
|
||||
``package_ensure``
|
||||
(optional) The state of the package
|
||||
Defaults to present
|
||||
|
||||
``bind_host``
|
||||
(optional) The cinder api bind address
|
||||
Defaults to 0.0.0.0
|
||||
|
||||
``ratelimits``
|
||||
(optional) The state of the service
|
||||
Defaults to undef. If undefined the default ratelimiting values are used.
|
||||
|
||||
``ratelimits_factory``
|
||||
(optional) Factory to use for ratelimiting
|
||||
Defaults to 'cinder.api.v1.limits:RateLimitingMiddleware.factory'
|
||||
|
||||
``default_volume_type``
|
||||
(optional) default volume type to use.
|
||||
This should contain the name of the default volume type to use.
|
||||
If not configured, it produces an error when creating a volume
|
||||
without specifying a type.
|
||||
Defaults to 'false'.
|
||||
|
||||
``validate``
|
||||
(optional) Whether to validate the service is working after any service refreshes
|
||||
Defaults to false
|
||||
|
||||
``validation_options``
|
||||
(optional) Service validation options
|
||||
Should be a hash of options defined in openstacklib::service_validation
|
||||
If empty, defaults values are taken from openstacklib function.
|
||||
Default command list volumes.
|
||||
Require validate set at True.
|
||||
Example:
|
||||
glance::api::validation_options:
|
||||
glance-api:
|
||||
command: check_cinder-api.py
|
||||
path: /usr/bin:/bin:/usr/sbin:/sbin
|
||||
provider: shell
|
||||
tries: 5
|
||||
try_sleep: 10
|
||||
Defaults to {}
|
@ -1,12 +0,0 @@
|
||||
class {'cinder::api':
|
||||
enabled => false,
|
||||
package_ensure => 'absent',
|
||||
keystone_password => 'not important as removed',
|
||||
}
|
||||
|
||||
include cinder::params
|
||||
|
||||
package { 'cinder':
|
||||
ensure => 'absent',
|
||||
name => $::cinder::params::package_name,
|
||||
}
|
@ -1,52 +0,0 @@
|
||||
$resource = hiera($::resource_name)
|
||||
|
||||
$keystone_password = $resource['input']['keystone_password']
|
||||
$keystone_enabled = $resource['input']['keystone_enabled']
|
||||
$keystone_tenant = $resource['input']['keystone_tenant']
|
||||
$keystone_user = $resource['input']['keystone_user']
|
||||
$keystone_auth_host = $resource['input']['keystone_auth_host']
|
||||
$keystone_auth_port = $resource['input']['keystone_auth_port']
|
||||
$keystone_auth_protocol = $resource['input']['keystone_auth_protocol']
|
||||
$keystone_auth_admin_prefix = $resource['input']['keystone_auth_admin_prefix']
|
||||
$keystone_auth_uri = $resource['input']['keystone_auth_uri']
|
||||
$os_region_name = $resource['input']['os_region_name']
|
||||
$service_port = $resource['input']['service_port']
|
||||
$service_workers = $resource['input']['service_workers']
|
||||
$package_ensure = $resource['input']['package_ensure']
|
||||
$bind_host = $resource['input']['bind_host']
|
||||
$ratelimits = $resource['input']['ratelimits']
|
||||
$default_volume_type = $resource['input']['default_volume_type']
|
||||
$ratelimits_factory = $resource['input']['ratelimits_factory']
|
||||
$validate = $resource['input']['validate']
|
||||
$validation_options = $resource['input']['validation_options']
|
||||
|
||||
include cinder::params
|
||||
|
||||
package { 'cinder':
|
||||
ensure => $package_ensure,
|
||||
name => $::cinder::params::package_name,
|
||||
} ->
|
||||
|
||||
class {'cinder::api':
|
||||
keystone_password => $keystone_password,
|
||||
keystone_enabled => $keystone_enabled,
|
||||
keystone_tenant => $keystone_tenant,
|
||||
keystone_user => $keystone_user,
|
||||
keystone_auth_host => $keystone_auth_host,
|
||||
keystone_auth_port => $keystone_auth_port,
|
||||
keystone_auth_protocol => $keystone_auth_protocol,
|
||||
keystone_auth_admin_prefix => $keystone_auth_admin_prefix,
|
||||
keystone_auth_uri => $keystone_auth_uri,
|
||||
os_region_name => $os_region_name,
|
||||
service_port => $service_port,
|
||||
service_workers => $service_workers,
|
||||
package_ensure => $package_ensure,
|
||||
bind_host => $bind_host,
|
||||
enabled => true,
|
||||
manage_service => true,
|
||||
ratelimits => $ratelimits,
|
||||
default_volume_type => $default_volume_type,
|
||||
ratelimits_factory => $ratelimits_factory,
|
||||
validate => $validate,
|
||||
validation_options => $validation_options,
|
||||
}
|
@ -1,56 +0,0 @@
|
||||
$resource = hiera($::resource_name)
|
||||
|
||||
$keystone_password = $resource['input']['keystone_password']
|
||||
$keystone_enabled = $resource['input']['keystone_enabled']
|
||||
$keystone_tenant = $resource['input']['keystone_tenant']
|
||||
$keystone_user = $resource['input']['keystone_user']
|
||||
$keystone_auth_host = $resource['input']['keystone_auth_host']
|
||||
$keystone_auth_port = $resource['input']['keystone_auth_port']
|
||||
$keystone_auth_protocol = $resource['input']['keystone_auth_protocol']
|
||||
$keystone_auth_admin_prefix = $resource['input']['keystone_auth_admin_prefix']
|
||||
$keystone_auth_uri = $resource['input']['keystone_auth_uri']
|
||||
$os_region_name = $resource['input']['os_region_name']
|
||||
$service_port = $resource['input']['service_port']
|
||||
$service_workers = $resource['input']['service_workers']
|
||||
$package_ensure = $resource['input']['package_ensure']
|
||||
$bind_host = $resource['input']['bind_host']
|
||||
$ratelimits = $resource['input']['ratelimits']
|
||||
$default_volume_type = $resource['input']['default_volume_type']
|
||||
$ratelimits_factory = $resource['input']['ratelimits_factory']
|
||||
$validate = $resource['input']['validate']
|
||||
$validation_options = $resource['input']['validation_options']
|
||||
|
||||
include cinder::params
|
||||
|
||||
package { 'cinder':
|
||||
ensure => $package_ensure,
|
||||
name => $::cinder::params::package_name,
|
||||
} ->
|
||||
|
||||
class {'cinder::api':
|
||||
keystone_password => $keystone_password,
|
||||
keystone_enabled => $keystone_enabled,
|
||||
keystone_tenant => $keystone_tenant,
|
||||
keystone_user => $keystone_user,
|
||||
keystone_auth_host => $keystone_auth_host,
|
||||
keystone_auth_port => $keystone_auth_port,
|
||||
keystone_auth_protocol => $keystone_auth_protocol,
|
||||
keystone_auth_admin_prefix => $keystone_auth_admin_prefix,
|
||||
keystone_auth_uri => $keystone_auth_uri,
|
||||
os_region_name => $os_region_name,
|
||||
service_port => $service_port,
|
||||
service_workers => $service_workers,
|
||||
package_ensure => $package_ensure,
|
||||
bind_host => $bind_host,
|
||||
enabled => true,
|
||||
manage_service => true,
|
||||
ratelimits => $ratelimits,
|
||||
default_volume_type => $default_volume_type,
|
||||
ratelimits_factory => $ratelimits_factory,
|
||||
validate => $validate,
|
||||
validation_options => $validation_options,
|
||||
}
|
||||
|
||||
notify { "restart cinder api":
|
||||
notify => Service["cinder-api"],
|
||||
}
|
@ -1,76 +0,0 @@
|
||||
handler: puppet
|
||||
version: 1.0.0
|
||||
input:
|
||||
keystone_password:
|
||||
schema: str!
|
||||
value: 'keystone'
|
||||
keystone_enabled:
|
||||
schema: bool
|
||||
value: true
|
||||
keystone_tenant:
|
||||
schema: str
|
||||
value: 'services'
|
||||
keystone_user:
|
||||
schema: str
|
||||
value: 'cinder'
|
||||
keystone_auth_host:
|
||||
schema: str
|
||||
value: 'localhost'
|
||||
keystone_auth_port:
|
||||
schema: int
|
||||
value: 35357
|
||||
keystone_auth_protocol:
|
||||
schema: str
|
||||
value: 'http'
|
||||
keystone_auth_admin_prefix:
|
||||
schema: bool
|
||||
value: false
|
||||
keystone_auth_uri:
|
||||
schema: bool
|
||||
value: false
|
||||
os_region_name:
|
||||
schema: str
|
||||
value:
|
||||
service_port:
|
||||
schema: int
|
||||
value: 5000
|
||||
service_workers:
|
||||
schema: int
|
||||
value: 1
|
||||
package_ensure:
|
||||
schema: str
|
||||
value: 'present'
|
||||
bind_host:
|
||||
schema: str
|
||||
value: '0.0.0.0'
|
||||
ratelimits:
|
||||
schema: str
|
||||
value:
|
||||
default_volume_type:
|
||||
schema: bool
|
||||
value: false
|
||||
ratelimits_factory:
|
||||
schema: str
|
||||
value: 'cinder.api.v1.limits:RateLimitingMiddleware.factory'
|
||||
validate:
|
||||
schema: bool
|
||||
value: false
|
||||
validation_options:
|
||||
schema: {}
|
||||
value: {}
|
||||
|
||||
git:
|
||||
schema: {repository: str!, branch: str!}
|
||||
value: {repository: 'https://github.com/openstack/puppet-cinder', branch: '5.1.0'}
|
||||
|
||||
ip:
|
||||
schema: str!
|
||||
value:
|
||||
# ssh_key:
|
||||
# schema: str!
|
||||
# value:
|
||||
# ssh_user:
|
||||
# schema: str!
|
||||
# value:
|
||||
|
||||
tags: [resource/cinder_api_service, resources/cinder_api, resources/cinder]
|
@ -1,10 +0,0 @@
|
||||
import requests
|
||||
|
||||
from solar.core.log import log
|
||||
|
||||
|
||||
def test(resource):
|
||||
log.debug('Testing cinder_api_puppet')
|
||||
requests.get(
|
||||
'http://%s:%s' % (resource.args['ip'], resource.args['service_port'])
|
||||
)
|
@ -1,39 +0,0 @@
|
||||
# Cinder Volume resource for puppet handler
|
||||
|
||||
Glance drive Cinder as a block storage backend to store image data.
|
||||
|
||||
# Parameters
|
||||
|
||||
source https://github.com/openstack/puppet-cinder/blob/5.1.0/manifests/glance.pp
|
||||
|
||||
``glance_api_servers``
|
||||
(optional) A list of the glance api servers available to cinder.
|
||||
Should be an array with [hostname|ip]:port
|
||||
Defaults to undef
|
||||
Note: for this resource, it is decomposed to *_host and *_port due to
|
||||
existing implementation limitations
|
||||
|
||||
``glance_api_version``
|
||||
(optional) Glance API version.
|
||||
Should be 1 or 2
|
||||
Defaults to 2 (current version)
|
||||
|
||||
``glance_num_retries``
|
||||
(optional) Number retries when downloading an image from glance.
|
||||
Defaults to 0
|
||||
|
||||
``glance_api_insecure``
|
||||
(optional) Allow to perform insecure SSL (https) requests to glance.
|
||||
Defaults to false
|
||||
|
||||
``glance_api_ssl_compression``
|
||||
(optional) Whether to attempt to negotiate SSL layer compression when
|
||||
using SSL (https) requests. Set to False to disable SSL
|
||||
layer compression. In some cases disabling this may improve
|
||||
data throughput, eg when high network bandwidth is available
|
||||
and you are using already compressed image formats such as qcow2.
|
||||
Defaults to false
|
||||
|
||||
``glance_request_timeout``
|
||||
(optional) http/https timeout value for glance operations.
|
||||
Defaults to undef
|
@ -1 +0,0 @@
|
||||
notify { 'Nothing to remove here': }
|
@ -1,18 +0,0 @@
|
||||
$resource = hiera($::resource_name)
|
||||
|
||||
$glance_api_version = $resource['input']['glance_api_version']
|
||||
$glance_num_retries = $resource['input']['glance_num_retries']
|
||||
$glance_api_insecure = $resource['input']['glance_api_insecure']
|
||||
$glance_api_ssl_compression = $resource['input']['glance_api_ssl_compression']
|
||||
$glance_request_timeout = $resource['input']['glance_request_timeout']
|
||||
$glance_api_servers_host = $resource['input']['glance_api_servers_host']
|
||||
$glance_api_servers_port = $resource['input']['glance_api_servers_port']
|
||||
|
||||
class {'cinder::glance':
|
||||
glance_api_servers => "${glance_api_servers_host}:${glance_api_servers_port}",
|
||||
glance_api_version => $glance_api_version,
|
||||
glance_num_retries => $glance_num_retries,
|
||||
glance_api_insecure => $glance_api_insecure,
|
||||
glance_api_ssl_compression => $glance_api_ssl_compression,
|
||||
glance_request_timeout => $glance_request_timeout,
|
||||
}
|
@ -1,41 +0,0 @@
|
||||
handler: puppet
|
||||
version: 1.0.0
|
||||
input:
|
||||
glance_api_version:
|
||||
schema: int
|
||||
value: 2
|
||||
glance_num_retries:
|
||||
schema: int
|
||||
value: 0
|
||||
glance_api_insecure:
|
||||
schema: bool
|
||||
value: false
|
||||
glance_api_ssl_compression:
|
||||
schema: bool
|
||||
value: false
|
||||
glance_request_timeout:
|
||||
schema: str
|
||||
value:
|
||||
|
||||
git:
|
||||
schema: {repository: str!, branch: str!}
|
||||
value: {repository: 'https://github.com/openstack/puppet-cinder', branch: '5.1.0'}
|
||||
|
||||
ip:
|
||||
schema: str!
|
||||
value:
|
||||
# ssh_key:
|
||||
# schema: str!
|
||||
# value:
|
||||
# ssh_user:
|
||||
# schema: str!
|
||||
# value:
|
||||
|
||||
glance_api_servers_port:
|
||||
schema: int
|
||||
value: 9292
|
||||
glance_api_servers_host:
|
||||
schema: 'str'
|
||||
value: 'localhost'
|
||||
|
||||
tags: [resource/cinder_glance_service, resources/cinder_glance, resources/cinder]
|
@ -1,112 +0,0 @@
|
||||
# Cinder resource for puppet handler
|
||||
|
||||
Controls a live cycle of the cinder entities,
|
||||
like the main puppet class, auth, DB, AMQP, packages,
|
||||
keystone user, role and endpoint.
|
||||
|
||||
# Parameters
|
||||
|
||||
source https://github.com/openstack/puppet-cinder/blob/5.1.0/manifests/init.pp
|
||||
|
||||
``database_connection``
|
||||
Url used to connect to database.
|
||||
(Optional) Defaults to
|
||||
'sqlite:////var/lib/cinder/cinder.sqlite'
|
||||
|
||||
``database_idle_timeout``
|
||||
Timeout when db connections should be reaped.
|
||||
(Optional) Defaults to 3600.
|
||||
|
||||
``database_min_pool_size``
|
||||
Minimum number of SQL connections to keep open in a pool.
|
||||
(Optional) Defaults to 1.
|
||||
|
||||
``database_max_pool_size``
|
||||
Maximum number of SQL connections to keep open in a pool.
|
||||
(Optional) Defaults to undef.
|
||||
|
||||
``database_max_retries``
|
||||
Maximum db connection retries during startup.
|
||||
Setting -1 implies an infinite retry count.
|
||||
(Optional) Defaults to 10.
|
||||
|
||||
``database_retry_interval``
|
||||
Interval between retries of opening a sql connection.
|
||||
(Optional) Defaults to 10.
|
||||
|
||||
``database_max_overflow``
|
||||
If set, use this value for max_overflow with sqlalchemy.
|
||||
(Optional) Defaults to undef.
|
||||
|
||||
``rabbit_use_ssl``
|
||||
(optional) Connect over SSL for RabbitMQ
|
||||
Defaults to false
|
||||
|
||||
``kombu_ssl_ca_certs``
|
||||
(optional) SSL certification authority file (valid only if SSL enabled).
|
||||
Defaults to undef
|
||||
|
||||
``kombu_ssl_certfile``
|
||||
(optional) SSL cert file (valid only if SSL enabled).
|
||||
Defaults to undef
|
||||
|
||||
``kombu_ssl_keyfile``
|
||||
(optional) SSL key file (valid only if SSL enabled).
|
||||
Defaults to undef
|
||||
|
||||
``kombu_ssl_version``
|
||||
(optional) SSL version to use (valid only if SSL enabled).
|
||||
Valid values are TLSv1, SSLv23 and SSLv3. SSLv2 may be
|
||||
available on some distributions.
|
||||
Defaults to 'TLSv1'
|
||||
|
||||
``amqp_durable_queues``
|
||||
Use durable queues in amqp.
|
||||
(Optional) Defaults to false.
|
||||
|
||||
``use_syslog``
|
||||
Use syslog for logging.
|
||||
(Optional) Defaults to false.
|
||||
|
||||
``log_facility``
|
||||
Syslog facility to receive log lines.
|
||||
(Optional) Defaults to LOG_USER.
|
||||
|
||||
``log_dir``
|
||||
(optional) Directory where logs should be stored.
|
||||
If set to boolean false, it will not log to any directory.
|
||||
Defaults to '/var/log/cinder'
|
||||
|
||||
``use_ssl``
|
||||
(optional) Enable SSL on the API server
|
||||
Defaults to false, not set
|
||||
|
||||
``cert_file``
|
||||
(optinal) Certificate file to use when starting API server securely
|
||||
Defaults to false, not set
|
||||
|
||||
``key_file``
|
||||
(optional) Private key file to use when starting API server securely
|
||||
Defaults to false, not set
|
||||
|
||||
``ca_file``
|
||||
(optional) CA certificate file to use to verify connecting clients
|
||||
Defaults to false, not set_
|
||||
|
||||
``mysql_module``
|
||||
(optional) Deprecated. Does nothing.
|
||||
|
||||
``storage_availability_zone``
|
||||
(optional) Availability zone of the node.
|
||||
Defaults to 'nova'
|
||||
|
||||
``default_availability_zone``
|
||||
(optional) Default availability zone for new volumes.
|
||||
If not set, the storage_availability_zone option value is used as
|
||||
the default for new volumes.
|
||||
Defaults to false
|
||||
|
||||
``sql_connection``
|
||||
DEPRECATED
|
||||
``sql_idle_timeout``
|
||||
DEPRECATED
|
@ -1,4 +0,0 @@
|
||||
class {'cinder':
|
||||
package_ensure => 'absent',
|
||||
rabbit_password => 'not important as removed',
|
||||
}
|
@ -1,116 +0,0 @@
|
||||
$resource = hiera($::resource_name)
|
||||
|
||||
$ip = $resource['input']['ip']
|
||||
|
||||
$db_user = $resource['input']['db_user']
|
||||
$db_password = $resource['input']['db_password']
|
||||
$db_name = $resource['input']['db_name']
|
||||
$db_host = $resource['input']['db_host']
|
||||
$db_port = $resource['input']['db_port']
|
||||
|
||||
$database_connection = $resource['input']['database_connection']
|
||||
$database_idle_timeout = $resource['input']['database_idle_timeout']
|
||||
$database_min_pool_size = $resource['input']['database_min_pool_size']
|
||||
$database_max_pool_size = $resource['input']['database_max_pool_size']
|
||||
$database_max_retries = $resource['input']['database_max_retries']
|
||||
$database_retry_interval = $resource['input']['database_retry_interval']
|
||||
$database_max_overflow = $resource['input']['database_max_overflow']
|
||||
$rpc_backend = $resource['input']['rpc_backend']
|
||||
$control_exchange = $resource['input']['control_exchange']
|
||||
$rabbit_host = $resource['input']['rabbit_host']
|
||||
$rabbit_port = $resource['input']['rabbit_port']
|
||||
$rabbit_hosts = $resource['input']['rabbit_hosts']
|
||||
$rabbit_virtual_host = $resource['input']['rabbit_virtual_host']
|
||||
$rabbit_userid = $resource['input']['rabbit_userid']
|
||||
$rabbit_password = $resource['input']['rabbit_password']
|
||||
$rabbit_use_ssl = $resource['input']['rabbit_use_ssl']
|
||||
$kombu_ssl_ca_certs = $resource['input']['kombu_ssl_ca_certs']
|
||||
$kombu_ssl_certfile = $resource['input']['kombu_ssl_certfile']
|
||||
$kombu_ssl_keyfile = $resource['input']['kombu_ssl_keyfile']
|
||||
$kombu_ssl_version = $resource['input']['kombu_ssl_version']
|
||||
$amqp_durable_queues = $resource['input']['amqp_durable_queues']
|
||||
$qpid_hostname = $resource['input']['qpid_hostname']
|
||||
$qpid_port = $resource['input']['qpid_port']
|
||||
$qpid_username = $resource['input']['qpid_username']
|
||||
$qpid_password = $resource['input']['qpid_password']
|
||||
$qpid_sasl_mechanisms = $resource['input']['qpid_sasl_mechanisms']
|
||||
$qpid_reconnect = $resource['input']['qpid_reconnect']
|
||||
$qpid_reconnect_timeout = $resource['input']['qpid_reconnect_timeout']
|
||||
$qpid_reconnect_limit = $resource['input']['qpid_reconnect_limit']
|
||||
$qpid_reconnect_interval_min = $resource['input']['qpid_reconnect_interval_min']
|
||||
$qpid_reconnect_interval_max = $resource['input']['qpid_reconnect_interval_max']
|
||||
$qpid_reconnect_interval = $resource['input']['qpid_reconnect_interval']
|
||||
$qpid_heartbeat = $resource['input']['qpid_heartbeat']
|
||||
$qpid_protocol = $resource['input']['qpid_protocol']
|
||||
$qpid_tcp_nodelay = $resource['input']['qpid_tcp_nodelay']
|
||||
$package_ensure = $resource['input']['package_ensure']
|
||||
$use_ssl = $resource['input']['use_ssl']
|
||||
$ca_file = $resource['input']['ca_file']
|
||||
$cert_file = $resource['input']['cert_file']
|
||||
$key_file = $resource['input']['key_file']
|
||||
$api_paste_config = $resource['input']['api_paste_config']
|
||||
$use_syslog = $resource['input']['use_syslog']
|
||||
$log_facility = $resource['input']['log_facility']
|
||||
$log_dir = $resource['input']['log_dir']
|
||||
$verbose = $resource['input']['verbose']
|
||||
$debug = $resource['input']['debug']
|
||||
$storage_availability_zone = $resource['input']['storage_availability_zone']
|
||||
$default_availability_zone = $resource['input']['default_availability_zone']
|
||||
$mysql_module = $resource['input']['mysql_module']
|
||||
# Do not apply the legacy stuff
|
||||
#$sql_connection = $resource['input']['sql_connection']
|
||||
$sql_idle_timeout = $resource['input']['sql_idle_timeout']
|
||||
|
||||
class {'cinder':
|
||||
database_connection => "mysql://${db_user}:${db_password}@${db_host}:${db_port}/${db_name}",
|
||||
database_idle_timeout => $database_idle_timeout,
|
||||
database_min_pool_size => $database_min_pool_size,
|
||||
database_max_pool_size => $database_max_pool_size,
|
||||
database_max_retries => $database_max_retries,
|
||||
database_retry_interval => $database_retry_interval,
|
||||
database_max_overflow => $database_max_overflow,
|
||||
rpc_backend => $rpc_backend,
|
||||
control_exchange => $control_exchange,
|
||||
rabbit_host => $rabbit_host,
|
||||
rabbit_port => $rabbit_port,
|
||||
rabbit_hosts => $rabbit_hosts,
|
||||
rabbit_virtual_host => $rabbit_virtual_host,
|
||||
rabbit_userid => $rabbit_userid,
|
||||
rabbit_password => $rabbit_password,
|
||||
rabbit_use_ssl => $rabbit_use_ssl,
|
||||
kombu_ssl_ca_certs => $kombu_ssl_ca_certs,
|
||||
kombu_ssl_certfile => $kombu_ssl_certfile,
|
||||
kombu_ssl_keyfile => $kombu_ssl_keyfile,
|
||||
kombu_ssl_version => $kombu_ssl_version,
|
||||
amqp_durable_queues => $amqp_durable_queues,
|
||||
qpid_hostname => $qpid_hostname,
|
||||
qpid_port => $qpid_port,
|
||||
qpid_username => $qpid_username,
|
||||
qpid_password => $qpid_password,
|
||||
qpid_sasl_mechanisms => $qpid_sasl_mechanisms,
|
||||
qpid_reconnect => $qpid_reconnect,
|
||||
qpid_reconnect_timeout => $qpid_reconnect_timeout,
|
||||
qpid_reconnect_limit => $qpid_reconnect_limit,
|
||||
qpid_reconnect_interval_min => $qpid_reconnect_interval_min,
|
||||
qpid_reconnect_interval_max => $qpid_reconnect_interval_max,
|
||||
qpid_reconnect_interval => $qpid_reconnect_interval,
|
||||
qpid_heartbeat => $qpid_heartbeat,
|
||||
qpid_protocol => $qpid_protocol,
|
||||
qpid_tcp_nodelay => $qpid_tcp_nodelay,
|
||||
package_ensure => $package_ensure,
|
||||
use_ssl => $use_ssl,
|
||||
ca_file => $ca_file,
|
||||
cert_file => $cert_file,
|
||||
key_file => $key_file,
|
||||
api_paste_config => $api_paste_config,
|
||||
use_syslog => $use_syslog,
|
||||
log_facility => $log_facility,
|
||||
log_dir => $log_dir,
|
||||
verbose => $verbose,
|
||||
debug => $debug,
|
||||
storage_availability_zone => $storage_availability_zone,
|
||||
default_availability_zone => $default_availability_zone,
|
||||
mysql_module => $mysql_module,
|
||||
sql_connection => $sql_connection,
|
||||
sql_idle_timeout => $sql_idle_timeout,
|
||||
}
|
@ -1,215 +0,0 @@
|
||||
handler: puppet
|
||||
actions:
|
||||
run: run.pp
|
||||
update: run.pp
|
||||
version: 1.0.0
|
||||
input:
|
||||
database_connection:
|
||||
schema: str
|
||||
value: 'sqlite:////var/lib/cinder/cinder.sqlite'
|
||||
database_idle_timeout:
|
||||
schema: int
|
||||
value: 3600
|
||||
database_min_pool_size:
|
||||
schema: int
|
||||
value: 1
|
||||
database_max_pool_size:
|
||||
schema: str
|
||||
value:
|
||||
database_max_retries:
|
||||
schema: int
|
||||
value: 10
|
||||
database_retry_interval:
|
||||
schema: int
|
||||
value: 10
|
||||
database_max_overflow:
|
||||
schema: str
|
||||
value:
|
||||
rpc_backend:
|
||||
schema: str
|
||||
value: 'cinder.openstack.common.rpc.impl_kombu'
|
||||
control_exchange:
|
||||
schema: str
|
||||
value: 'openstack'
|
||||
rabbit_host:
|
||||
schema: str
|
||||
value: '127.0.0.1'
|
||||
rabbit_port:
|
||||
schema: int
|
||||
value: 5672
|
||||
rabbit_hosts:
|
||||
schema: bool
|
||||
value: false
|
||||
rabbit_virtual_host:
|
||||
schema: str
|
||||
value: '/'
|
||||
rabbit_userid:
|
||||
schema: str
|
||||
value: 'guest'
|
||||
rabbit_password:
|
||||
schema: str!
|
||||
value: 'rabbit'
|
||||
rabbit_use_ssl:
|
||||
schema: bool
|
||||
value: false
|
||||
kombu_ssl_ca_certs:
|
||||
schema: str
|
||||
value:
|
||||
kombu_ssl_certfile:
|
||||
schema: str
|
||||
value:
|
||||
kombu_ssl_keyfile:
|
||||
schema: str
|
||||
value:
|
||||
kombu_ssl_version:
|
||||
schema: str
|
||||
value: 'TLSv1'
|
||||
amqp_durable_queues:
|
||||
schema: bool
|
||||
value: false
|
||||
qpid_hostname:
|
||||
schema: str
|
||||
value: 'localhost'
|
||||
qpid_port:
|
||||
schema: int
|
||||
value: 5672
|
||||
qpid_username:
|
||||
schema: str
|
||||
value: 'guest'
|
||||
qpid_password:
|
||||
schema: str!
|
||||
value: 'qpid'
|
||||
qpid_sasl_mechanisms:
|
||||
schema: bool
|
||||
value: false
|
||||
qpid_reconnect:
|
||||
schema: bool
|
||||
value: true
|
||||
qpid_reconnect_timeout:
|
||||
schema: int
|
||||
value: 0
|
||||
qpid_reconnect_limit:
|
||||
schema: int
|
||||
value: 0
|
||||
qpid_reconnect_interval_min:
|
||||
schema: int
|
||||
value: 0
|
||||
qpid_reconnect_interval_max:
|
||||
schema: int
|
||||
value: 0
|
||||
qpid_reconnect_interval:
|
||||
schema: int
|
||||
value: 0
|
||||
qpid_heartbeat:
|
||||
schema: int
|
||||
value: 60
|
||||
qpid_protocol:
|
||||
schema: str
|
||||
value: 'tcp'
|
||||
qpid_tcp_nodelay:
|
||||
schema: bool
|
||||
value: true
|
||||
package_ensure:
|
||||
schema: str
|
||||
value: 'present'
|
||||
use_ssl:
|
||||
schema: bool
|
||||
value: false
|
||||
ca_file:
|
||||
schema: bool
|
||||
value: false
|
||||
cert_file:
|
||||
schema: bool
|
||||
value: false
|
||||
key_file:
|
||||
schema: bool
|
||||
value: false
|
||||
api_paste_config:
|
||||
schema: str
|
||||
value: '/etc/cinder/api-paste.ini'
|
||||
use_syslog:
|
||||
schema: bool
|
||||
value: false
|
||||
log_facility:
|
||||
schema: str
|
||||
value: 'LOG_USER'
|
||||
log_dir:
|
||||
schema: str
|
||||
value: '/var/log/cinder'
|
||||
verbose:
|
||||
schema: bool
|
||||
value: false
|
||||
debug:
|
||||
schema: bool
|
||||
value: false
|
||||
storage_availability_zone:
|
||||
schema: str
|
||||
value: 'nova'
|
||||
default_availability_zone:
|
||||
schema: bool
|
||||
value: false
|
||||
mysql_module:
|
||||
schema: str
|
||||
value:
|
||||
sql_connection:
|
||||
schema: str
|
||||
value:
|
||||
sql_idle_timeout:
|
||||
schema: str
|
||||
value:
|
||||
|
||||
db_user:
|
||||
schema: str!
|
||||
value: cinder
|
||||
db_password:
|
||||
schema: str!
|
||||
value: cinder
|
||||
db_name:
|
||||
schema: str!
|
||||
value: cinder
|
||||
db_host:
|
||||
schema: str!
|
||||
value:
|
||||
db_port:
|
||||
schema: int!
|
||||
value:
|
||||
|
||||
port:
|
||||
schema: int!
|
||||
value: 8776
|
||||
|
||||
module:
|
||||
schema: {name: str!, type: str, url: str, ref: str}
|
||||
value: {name: 'cinder', type: 'git', url: 'https://github.com/openstack/puppet-cinder', ref: '5.1.0'}
|
||||
|
||||
keystone_host:
|
||||
schema: str!
|
||||
value:
|
||||
keystone_port:
|
||||
schema: int!
|
||||
value:
|
||||
keystone_user:
|
||||
schema: str!
|
||||
value:
|
||||
keystone_password:
|
||||
schema: str!
|
||||
value:
|
||||
keystone_tenant:
|
||||
schema: str!
|
||||
value:
|
||||
|
||||
# forge:
|
||||
# schema: str!
|
||||
# value: 'stackforge-cinder'
|
||||
|
||||
ip:
|
||||
schema: str!
|
||||
value:
|
||||
# ssh_key:
|
||||
# schema: str!
|
||||
# value:
|
||||
# ssh_user:
|
||||
# schema: str!
|
||||
# value:
|
||||
|
||||
tags: [resource/cinder_service, resources/cinder]
|
@ -1,10 +0,0 @@
|
||||
import requests
|
||||
|
||||
from solar.core.log import log
|
||||
|
||||
|
||||
def test(resource):
|
||||
log.debug('Testing cinder_puppet')
|
||||
requests.get(
|
||||
'http://%s:%s' % (resource.args['ip'], resource.args['port'])
|
||||
)
|
@ -1,3 +0,0 @@
|
||||
# Cinder Scheduler resource for puppet handler
|
||||
|
||||
Setup and configure the cinder scheduler service
|
@ -1,4 +0,0 @@
|
||||
class {'cinder::scheduler':
|
||||
enabled => false,
|
||||
package_ensure => 'absent',
|
||||
}
|
@ -1,18 +0,0 @@
|
||||
$resource = hiera($::resource_name)
|
||||
|
||||
$scheduler_driver = $resource['input']['scheduler_driver']
|
||||
$package_ensure = $resource['input']['package_ensure']
|
||||
|
||||
include cinder::params
|
||||
|
||||
package { 'cinder':
|
||||
ensure => $package_ensure,
|
||||
name => $::cinder::params::package_name,
|
||||
} ->
|
||||
|
||||
class {'cinder::scheduler':
|
||||
scheduler_driver => $scheduler_driver,
|
||||
package_ensure => $package_ensure,
|
||||
enabled => true,
|
||||
manage_service => true,
|
||||
}
|
@ -1,22 +0,0 @@
|
||||
$resource = hiera($::resource_name)
|
||||
|
||||
$scheduler_driver = $resource['input']['scheduler_driver']
|
||||
$package_ensure = $resource['input']['package_ensure']
|
||||
|
||||
include cinder::params
|
||||
|
||||
package { 'cinder':
|
||||
ensure => $package_ensure,
|
||||
name => $::cinder::params::package_name,
|
||||
} ->
|
||||
|
||||
class {'cinder::scheduler':
|
||||
scheduler_driver => $scheduler_driver,
|
||||
package_ensure => $package_ensure,
|
||||
enabled => true,
|
||||
manage_service => true,
|
||||
}
|
||||
|
||||
notify { "restart cinder volume":
|
||||
notify => Service["cinder-scheduler"],
|
||||
}
|
@ -1,25 +0,0 @@
|
||||
handler: puppet
|
||||
version: 1.0.0
|
||||
input:
|
||||
scheduler_driver:
|
||||
schema: str
|
||||
value:
|
||||
package_ensure:
|
||||
schema: str
|
||||
value: 'present'
|
||||
|
||||
git:
|
||||
schema: {repository: str!, branch: str!}
|
||||
value: {repository: 'https://github.com/openstack/puppet-cinder', branch: '5.1.0'}
|
||||
|
||||
ip:
|
||||
schema: str!
|
||||
value:
|
||||
# ssh_key:
|
||||
# schema: str!
|
||||
# value:
|
||||
# ssh_user:
|
||||
# schema: str!
|
||||
# value:
|
||||
|
||||
tags: [resource/cinder_scheduler_service, resources/cinder_scheduler, resources/cinder]
|
@ -1,12 +0,0 @@
|
||||
import requests
|
||||
|
||||
from solar.core.log import log
|
||||
|
||||
|
||||
def test(resource):
|
||||
log.debug('Testing cinder_scheduler_puppet')
|
||||
# requests.get(
|
||||
# 'http://%s:%s' % (resource.args['ip'], resource.args['port'])
|
||||
# TODO(bogdando) figure out how to test this
|
||||
# http://docs.openstack.org/developer/nova/devref/scheduler.html
|
||||
# )
|
@ -1,8 +0,0 @@
|
||||
# Cinder Volume resource for puppet handler
|
||||
|
||||
Setup and configure the cinder volume service.
|
||||
Configure ISCSI volume backend for cinder volume nodes, if specified
|
||||
(Default true)
|
||||
|
||||
source https://github.com/openstack/puppet-cinder/blob/5.1.0/manifests/volume/iscsi.pp
|
||||
source https://github.com/openstack/puppet-cinder/blob/5.1.0/manifests/volume.pp
|
@ -1,4 +0,0 @@
|
||||
class {'cinder::volume':
|
||||
enabled => false,
|
||||
package_ensure => 'absent',
|
||||
}
|
@ -1,31 +0,0 @@
|
||||
$resource = hiera($::resource_name)
|
||||
|
||||
$package_ensure = $resource['input']['package_ensure']
|
||||
$use_iscsi_backend = $resource['input']['use_iscsi_backend']
|
||||
|
||||
$iscsi_ip_address = $resource['input']['iscsi_ip_address']
|
||||
$volume_driver = $resource['input']['volume_driver']
|
||||
$volume_group = $resource['input']['volume_group']
|
||||
$iscsi_helper = $resource['input']['iscsi_helper']
|
||||
|
||||
include cinder::params
|
||||
|
||||
package { 'cinder':
|
||||
ensure => $package_ensure,
|
||||
name => $::cinder::params::package_name,
|
||||
} ->
|
||||
|
||||
class {'cinder::volume':
|
||||
package_ensure => $package_ensure,
|
||||
enabled => true,
|
||||
manage_service => true,
|
||||
}
|
||||
|
||||
if $use_iscsi_backend {
|
||||
class {'cinder::volume::iscsi':
|
||||
iscsi_ip_address => $iscsi_ip_address,
|
||||
volume_driver => $volume_driver,
|
||||
volume_group => $volume_group,
|
||||
iscsi_helper => $iscsi_helper,
|
||||
}
|
||||
}
|
@ -1,26 +0,0 @@
|
||||
$resource = hiera($::resource_name)
|
||||
|
||||
$package_ensure = $resource['input']['package_ensure']
|
||||
$use_iscsi_backend = $resource['input']['use_iscsi_backend']
|
||||
|
||||
$iscsi_ip_address = $resource['input']['iscsi_ip_address']
|
||||
$volume_driver = $resource['input']['volume_driver']
|
||||
$volume_group = $resource['input']['volume_group']
|
||||
$iscsi_helper = $resource['input']['iscsi_helper']
|
||||
|
||||
include cinder::params
|
||||
|
||||
package { 'cinder':
|
||||
ensure => $package_ensure,
|
||||
name => $::cinder::params::package_name,
|
||||
} ->
|
||||
|
||||
class {'cinder::volume':
|
||||
package_ensure => $package_ensure,
|
||||
enabled => true,
|
||||
manage_service => true,
|
||||
}
|
||||
|
||||
notify { "restart cinder volume":
|
||||
notify => Service["cinder-volume"],
|
||||
}
|
@ -1,38 +0,0 @@
|
||||
handler: puppet
|
||||
version: 1.0.0
|
||||
input:
|
||||
package_ensure:
|
||||
schema: str
|
||||
value: 'present'
|
||||
iscsi_ip_address:
|
||||
schema: str
|
||||
value: '127.0.0.1'
|
||||
volume_driver:
|
||||
schema: str
|
||||
value: 'cinder.volume.drivers.lvm.LVMISCSIDriver'
|
||||
volume_group:
|
||||
schema: str
|
||||
value: 'cinder-volumes'
|
||||
iscsi_helper:
|
||||
schema: str
|
||||
value: 'tgtadm'
|
||||
|
||||
use_iscsi_backend:
|
||||
schema: bool
|
||||
value: true
|
||||
|
||||
git:
|
||||
schema: {repository: str!, branch: str!}
|
||||
value: {repository: 'https://github.com/openstack/puppet-cinder', branch: '5.1.0'}
|
||||
|
||||
ip:
|
||||
schema: str!
|
||||
value:
|
||||
# ssh_key:
|
||||
# schema: str!
|
||||
# value:
|
||||
# ssh_user:
|
||||
# schema: str!
|
||||
# value:
|
||||
|
||||
tags: [resource/cinder_volume_service, resources/cinder_volume, resources/cinder]
|
@ -1,12 +0,0 @@
|
||||
import requests
|
||||
|
||||
from solar.core.log import log
|
||||
|
||||
|
||||
def test(resource):
|
||||
log.debug('Testing cinder_volume_puppet')
|
||||
# requests.get(
|
||||
# 'http://%s:%s' % (resource.args['ip'], resource.args['port'])
|
||||
# TODO(bogdando) figure out how to test this
|
||||
# http://docs.openstack.org/developer/nova/devref/volume.html
|
||||
# )
|
@ -1,22 +0,0 @@
|
||||
- hosts: '*'
|
||||
sudo: yes
|
||||
gather_facts: false
|
||||
# this is default variables, they will be overwritten by resource one
|
||||
vars:
|
||||
networks:
|
||||
mgmt:
|
||||
address: 172.18.10.6
|
||||
bridge: br-test0
|
||||
bridge_address: 172.18.10.252/24
|
||||
interface: eth1
|
||||
netmask: 255.255.255.0
|
||||
type: veth
|
||||
tasks:
|
||||
- shell: ip l add {{item.value.bridge}} type bridge
|
||||
with_dict: networks
|
||||
ignore_errors: true
|
||||
- shell: ip l set {{item.value.bridge}} up
|
||||
with_dict: networks
|
||||
- shell: ip a add dev {{item.value.bridge}} {{item.value.bridge_address}}
|
||||
with_dict: networks
|
||||
ignore_errors: true
|
@ -1,16 +0,0 @@
|
||||
handler: ansible_playbook
|
||||
version: 1.0.0
|
||||
actions:
|
||||
input:
|
||||
ip:
|
||||
schema: str!
|
||||
value:
|
||||
# ssh_key:
|
||||
# schema: str!
|
||||
# value:
|
||||
# ssh_user:
|
||||
# schema: str!
|
||||
# value:
|
||||
networks:
|
||||
schema: {}
|
||||
value:
|
@ -1,5 +0,0 @@
|
||||
|
||||
- hosts: [{{host}}]
|
||||
sudo: yes
|
||||
tasks:
|
||||
- shell: echo `/sbin/ifconfig`
|
@ -1,6 +0,0 @@
|
||||
|
||||
- hosts: [{{host}}]
|
||||
sudo: yes
|
||||
tasks:
|
||||
- shell: docker stop {{ resource_name }}
|
||||
- shell: docker rm {{ resource_name }}
|
@ -1,24 +0,0 @@
|
||||
- hosts: [{{host}}]
|
||||
sudo: yes
|
||||
tasks:
|
||||
- docker:
|
||||
name: {{ resource_name }}
|
||||
image: {{ image }}
|
||||
state: running
|
||||
net: host
|
||||
{% if ports.value %}
|
||||
ports:
|
||||
{% for port in ports.value %}
|
||||
- {{ port['value'] }}:{{ port['value'] }}
|
||||
{% endfor %}
|
||||
{% endif %}
|
||||
{% if host_binds.value %}
|
||||
volumes:
|
||||
# TODO: host_binds might need more work
|
||||
# Currently it's not that trivial to pass custom src: dst here
|
||||
# (when a config variable is passed here from other resource)
|
||||
# so we mount it to the same directory as on host
|
||||
{% for bind in host_binds.value %}
|
||||
- {{ bind['value']['src'] }}:{{ bind['value']['dst'] }}:{{ bind['value'].get('mode', 'ro') }}
|
||||
{% endfor %}
|
||||
{% endif %}
|
@ -1,12 +0,0 @@
|
||||
handler: ansible
|
||||
version: 1.0.0
|
||||
input:
|
||||
ip:
|
||||
type: str!
|
||||
value:
|
||||
image:
|
||||
type: str!
|
||||
value:
|
||||
export_volumes:
|
||||
type: str!
|
||||
value:
|
@ -1,6 +0,0 @@
|
||||
- hosts: [{{host}}]
|
||||
sudo: yes
|
||||
|
||||
tasks:
|
||||
- lineinfile: create=yes dest=/etc/dnsmasq.d/no_pxe_{{exclude_mac_pxe | replace(':', '_')}}.conf line="dhcp-host={{exclude_mac_pxe}},set:nopxe"
|
||||
- shell: service dnsmasq restart
|
@ -1,2 +0,0 @@
|
||||
- hosts: [{{host}}]
|
||||
sudo: yes
|
@ -1,17 +0,0 @@
|
||||
handler: ansible
|
||||
version: 1.0.0
|
||||
|
||||
actions:
|
||||
exclude_mac_pxe: exclude_mac_pxe.yaml
|
||||
run: run.yaml
|
||||
|
||||
input:
|
||||
ip:
|
||||
schema: str!
|
||||
value:
|
||||
|
||||
exclude_mac_pxe:
|
||||
schema: str!
|
||||
value:
|
||||
|
||||
tags: [resources=dnsmasq]
|
@ -1,9 +0,0 @@
|
||||
|
||||
- hosts: [{{host}}]
|
||||
sudo: yes
|
||||
tasks:
|
||||
- shell: docker --version
|
||||
ignore_errors: true
|
||||
register: docker_version
|
||||
- shell: curl -sSL https://get.docker.com/ | sudo sh
|
||||
when: docker_version|failed
|
@ -1,15 +0,0 @@
|
||||
handler: ansible
|
||||
version: 1.0.0
|
||||
|
||||
input:
|
||||
ip:
|
||||
schema: str!
|
||||
value:
|
||||
# ssh_user:
|
||||
# schema: str!
|
||||
# value:
|
||||
# ssh_key:
|
||||
# schema: str!
|
||||
# value:
|
||||
|
||||
tags: [resources/docker]
|
@ -1,6 +0,0 @@
|
||||
|
||||
- hosts: [{{host}}]
|
||||
sudo: yes
|
||||
tasks:
|
||||
- shell: docker stop {{ resource_name }}
|
||||
- shell: docker rm {{ resource_name }}
|
@ -1,37 +0,0 @@
|
||||
|
||||
- hosts: [{{host}}]
|
||||
sudo: yes
|
||||
tasks:
|
||||
- docker:
|
||||
name: {{ resource_name }}
|
||||
image: {{ image }}
|
||||
state: running
|
||||
net: host
|
||||
{% if ports %}
|
||||
ports:
|
||||
{% for port in ports %}
|
||||
- {{ port }}:{{ port }}
|
||||
{% endfor %}
|
||||
expose:
|
||||
{% for port in ports %}
|
||||
- {{ port }}
|
||||
{% endfor %}
|
||||
{% endif %}
|
||||
|
||||
{% if host_binds.value %}
|
||||
volumes:
|
||||
# TODO: host_binds might need more work
|
||||
# Currently it's not that trivial to pass custom src: dst here
|
||||
# (when a config variable is passed here from other resource)
|
||||
# so we mount it to the same directory as on host
|
||||
{% for bind in host_binds.value %}
|
||||
- {{ bind['value']['src'] }}:{{ bind['value']['dst'] }}:{{ bind['value'].get('mode', 'ro') }}
|
||||
{% endfor %}
|
||||
{% endif %}
|
||||
|
||||
{% if env %}
|
||||
env:
|
||||
{% for key, value in env.iteritems() %}
|
||||
{{ key }}: {{ value }}
|
||||
{% endfor %}
|
||||
{% endif %}
|
@ -1,37 +0,0 @@
|
||||
|
||||
- hosts: [{{host}}]
|
||||
sudo: yes
|
||||
tasks:
|
||||
- docker:
|
||||
name: {{ resource_name }}
|
||||
image: {{ image }}
|
||||
state: reloaded
|
||||
net: host
|
||||
{% if ports %}
|
||||
ports:
|
||||
{% for port in ports %}
|
||||
- {{ port }}:{{ port }}
|
||||
{% endfor %}
|
||||
expose:
|
||||
{% for port in ports %}
|
||||
- {{ port }}
|
||||
{% endfor %}
|
||||
{% endif %}
|
||||
|
||||
{% if host_binds.value %}
|
||||
volumes:
|
||||
# TODO: host_binds might need more work
|
||||
# Currently it's not that trivial to pass custom src: dst here
|
||||
# (when a config variable is passed here from other resource)
|
||||
# so we mount it to the same directory as on host
|
||||
{% for bind in host_binds.value %}
|
||||
- {{ bind['value']['src'] }}:{{ bind['value']['dst'] }}:{{ bind['value'].get('mode', 'ro') }}
|
||||
{% endfor %}
|
||||
{% endif %}
|
||||
|
||||
{% if env %}
|
||||
env:
|
||||
{% for key, value in env.iteritems() %}
|
||||
{{ key }}: {{ value }}
|
||||
{% endfor %}
|
||||
{% endif %}
|
@ -1,29 +0,0 @@
|
||||
handler: ansible
|
||||
version: 1.0.0
|
||||
input:
|
||||
ip:
|
||||
schema: str!
|
||||
value:
|
||||
image:
|
||||
schema: str!
|
||||
value:
|
||||
ports:
|
||||
schema: [int]
|
||||
value: []
|
||||
host_binds:
|
||||
schema: [{value: {src: str, dst: str, mode: str}}]
|
||||
value: []
|
||||
volume_binds:
|
||||
schema: [{src: str, dst: str, mode: str}]
|
||||
value: []
|
||||
env:
|
||||
schema: {}
|
||||
value: {}
|
||||
# ssh_user:
|
||||
# schema: str!
|
||||
# value: []
|
||||
# ssh_key:
|
||||
# schema: str!
|
||||
# value: []
|
||||
|
||||
tags: [resource/container]
|
@ -1,10 +0,0 @@
|
||||
#!/usr/bin/env python
|
||||
|
||||
import sys
|
||||
import json
|
||||
|
||||
data = json.loads(sys.stdin.read())
|
||||
|
||||
rst = {'val_x_val': int(data['val'])**2}
|
||||
|
||||
sys.stdout.write(json.dumps(rst))
|
@ -1,11 +0,0 @@
|
||||
handler: none
|
||||
version: 1.0.0
|
||||
managers:
|
||||
- managers/manager.py
|
||||
input:
|
||||
val:
|
||||
schema: int!
|
||||
value: 2
|
||||
val_x_val:
|
||||
schema: int
|
||||
value:
|
@ -1,3 +0,0 @@
|
||||
#!/bin/bash
|
||||
|
||||
rm {{ path }}
|
@ -1,3 +0,0 @@
|
||||
#!/bin/bash
|
||||
|
||||
touch {{ path }}
|
Some files were not shown because too many files have changed in this diff Show More
Loading…
Reference in New Issue
Block a user