Created initial data for Tuskar
This patch adds initial data to Tuskar in the form of four Overcloud Roles (controller, computer, object_storage, and block_storage). Change-Id: Ie3ea02333144b2611025e7b2d84bf23d986c5069
This commit is contained in:
parent
ce32ea4877
commit
71201df8e7
@ -91,14 +91,21 @@ output) with::
|
||||
|
||||
$ sqlite3 tuskar/openstack/common/db/tuskar.sqlite .schema
|
||||
|
||||
Then, launch the app and try curl to see if it works::
|
||||
Then, launch the app::
|
||||
|
||||
$ tuskar-api --config-file etc/tuskar/tuskar.conf
|
||||
$ curl -v -X GET -H 'Accept: application/json' http://0.0.0.0:8585/v1/overclouds/ | python -mjson.tool
|
||||
|
||||
Assuming this is your first time running with a new database, you should
|
||||
simply get '[]' back from curl above. Currently the api supports only
|
||||
json return type, so we request that in the example call.
|
||||
populate the DB with some initial data::
|
||||
|
||||
$ python tools/initial_data.py
|
||||
|
||||
This will create four Overcloud Roles. You can then verify that everything
|
||||
worked by running:
|
||||
|
||||
$ curl -v -X GET -H 'Accept: application/json' http://0.0.0.0:8585/v1/overcloud_roles/ | python -mjson.tool
|
||||
|
||||
This command should return JSON for four Overcloud Roles.
|
||||
|
||||
Running Tuskar API
|
||||
------------------
|
||||
|
138
tools/initial_data.py
Normal file
138
tools/initial_data.py
Normal file
@ -0,0 +1,138 @@
|
||||
#!/usr/bin/env python
|
||||
|
||||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
# Using the REST API, populates the DB with some sample data
|
||||
# Based on python-ceilometerclient/ceilometerclient/common/http.py
|
||||
|
||||
import httplib
|
||||
import json
|
||||
import logging
|
||||
import socket
|
||||
import six.moves.urllib.parse as urlparse
|
||||
import uuid
|
||||
|
||||
|
||||
LOG = logging.getLogger(__name__)
|
||||
|
||||
|
||||
def log_curl_request(conn, base_url, url, method, kwargs):
|
||||
curl = ['curl -i -X %s' % method]
|
||||
|
||||
for (key, value) in kwargs['headers'].items():
|
||||
header = '-H \'%s: %s\'' % (key, value)
|
||||
curl.append(header)
|
||||
|
||||
if 'body' in kwargs:
|
||||
curl.append('-d \'%s\'' % kwargs['body'])
|
||||
|
||||
curl.append('http://%s:%d%s%s' % (conn.host, conn.port, base_url, url))
|
||||
LOG.debug(' '.join(curl))
|
||||
|
||||
|
||||
def log_http_response(resp, body=None):
|
||||
status = (resp.version / 10.0, resp.status, resp.reason)
|
||||
dump = ['\nHTTP/%.1f %s %s' % status]
|
||||
dump.extend(['%s: %s' % (k, v) for k, v in resp.getheaders()])
|
||||
dump.append('')
|
||||
if body:
|
||||
dump.extend([body, ''])
|
||||
LOG.debug('\n'.join(dump))
|
||||
|
||||
|
||||
def make_connection_url(base_url, url):
|
||||
return '%s/%s' % (base_url.rstrip('/'), url.lstrip('/'))
|
||||
|
||||
|
||||
def http_request(conn, base_url, url, method, **kwargs):
|
||||
log_curl_request(conn, base_url, url, method, kwargs)
|
||||
|
||||
try:
|
||||
conn_url = make_connection_url(base_url, url)
|
||||
conn.request(method, conn_url, **kwargs)
|
||||
resp = conn.getresponse()
|
||||
except socket.gaierror as e:
|
||||
message = ('Error finding address for %(url)s: %(e)s' %
|
||||
{'url': url, 'e': e})
|
||||
raise RuntimeError(message)
|
||||
except (socket.error, socket.timeout) as e:
|
||||
message = ('Error communicating with %(endpoint)s %(e)s' %
|
||||
{'endpoint': 'http://%s:%d' % (conn.host, conn.port),
|
||||
'e': e})
|
||||
raise RuntimeError(message)
|
||||
|
||||
body = resp.read()
|
||||
log_http_response(resp, body)
|
||||
|
||||
if 300 <= resp.status < 600:
|
||||
LOG.warn('Request returned failure/redirect status.')
|
||||
raise RuntimeError('Status code %d returned' % resp.status)
|
||||
|
||||
return resp, body
|
||||
|
||||
|
||||
def json_request(conn, base_url, url, method, **kwargs):
|
||||
kwargs.setdefault('headers', {})
|
||||
kwargs['headers'].setdefault('Content-Type', 'application/json')
|
||||
kwargs['headers'].setdefault('Accept', 'application/json')
|
||||
|
||||
if 'body' in kwargs:
|
||||
kwargs['body'] = json.dumps(kwargs['body'])
|
||||
|
||||
resp, body = http_request(conn, base_url, url, method, **kwargs)
|
||||
content_type = resp.getheader('content-type', None)
|
||||
|
||||
if resp.status == 204 or resp.status == 205 or content_type is None:
|
||||
body = None
|
||||
elif 'application/json' in content_type:
|
||||
try:
|
||||
body = json.loads(body)
|
||||
except ValueError:
|
||||
LOG.error('Could not decode response body as JSON')
|
||||
else:
|
||||
body = None
|
||||
|
||||
return resp, body
|
||||
|
||||
|
||||
def create_overcloud_role(conn, base_url, name, description, image_name):
|
||||
return json_request(conn, base_url, '/overcloud_roles', 'POST',
|
||||
body=dict(name=name, description=description,
|
||||
image_name=image_name))
|
||||
|
||||
|
||||
def generate_data():
|
||||
conn = httplib.HTTPConnection('localhost', 8585)
|
||||
base_url = '/v1'
|
||||
|
||||
create_overcloud_role(conn, base_url,
|
||||
name='Controller',
|
||||
description='controller role',
|
||||
image_name='overcloud-control')
|
||||
create_overcloud_role(conn, base_url,
|
||||
name='Compute',
|
||||
description='compute role',
|
||||
image_name='overcloud-compute')
|
||||
create_overcloud_role(conn, base_url,
|
||||
name='Object Storage',
|
||||
description='object storage role',
|
||||
image_name='overcloud-object-storage')
|
||||
create_overcloud_role(conn, base_url,
|
||||
name='Block Storage',
|
||||
description='block storage role',
|
||||
image_name='overcloud-block-storage')
|
||||
|
||||
|
||||
if __name__ == '__main__':
|
||||
logging.basicConfig(format='%(message)s', level=logging.DEBUG)
|
||||
generate_data()
|
@ -1,230 +0,0 @@
|
||||
#!/usr/bin/env python
|
||||
|
||||
# Copyright 2012 OpenStack LLC.
|
||||
# All Rights Reserved.
|
||||
# Copyright 2013 Red Hat, Inc.
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
# Using the REST API, populates the DB with some sample data
|
||||
# Based on python-ceilometerclient/ceilometerclient/common/http.py
|
||||
|
||||
import httplib
|
||||
import json
|
||||
import logging
|
||||
import socket
|
||||
import urlparse
|
||||
import uuid
|
||||
|
||||
LOG = logging.getLogger(__name__)
|
||||
|
||||
|
||||
def log_curl_request(conn, base_url, url, method, kwargs):
|
||||
curl = ['curl -i -X %s' % method]
|
||||
|
||||
for (key, value) in kwargs['headers'].items():
|
||||
header = '-H \'%s: %s\'' % (key, value)
|
||||
curl.append(header)
|
||||
|
||||
if 'body' in kwargs:
|
||||
curl.append('-d \'%s\'' % kwargs['body'])
|
||||
|
||||
curl.append('http://%s:%d%s%s' % (conn.host, conn.port, base_url, url))
|
||||
LOG.debug(' '.join(curl))
|
||||
|
||||
|
||||
def log_http_response(resp, body=None):
|
||||
status = (resp.version / 10.0, resp.status, resp.reason)
|
||||
dump = ['\nHTTP/%.1f %s %s' % status]
|
||||
dump.extend(['%s: %s' % (k, v) for k, v in resp.getheaders()])
|
||||
dump.append('')
|
||||
if body:
|
||||
dump.extend([body, ''])
|
||||
LOG.debug('\n'.join(dump))
|
||||
|
||||
|
||||
def make_connection_url(base_url, url):
|
||||
return '%s/%s' % (base_url.rstrip('/'), url.lstrip('/'))
|
||||
|
||||
|
||||
def http_request(conn, base_url, url, method, **kwargs):
|
||||
log_curl_request(conn, base_url, url, method, kwargs)
|
||||
|
||||
try:
|
||||
conn_url = make_connection_url(base_url, url)
|
||||
conn.request(method, conn_url, **kwargs)
|
||||
resp = conn.getresponse()
|
||||
except socket.gaierror as e:
|
||||
message = ('Error finding address for %(url)s: %(e)s' %
|
||||
{'url': url, 'e': e})
|
||||
raise RuntimeError(message)
|
||||
except (socket.error, socket.timeout) as e:
|
||||
message = ('Error communicating with %(endpoint)s %(e)s' %
|
||||
{'endpoint': 'http://%s:%d' % (conn.host, conn.port),
|
||||
'e': e})
|
||||
raise RuntimeError(message)
|
||||
|
||||
if 300 <= resp.status < 600:
|
||||
LOG.warn('Request returned failure/redirect status.')
|
||||
raise RuntimeError('Status code %d returned' % resp.status)
|
||||
|
||||
body = resp.read()
|
||||
|
||||
log_http_response(resp, body)
|
||||
|
||||
return resp, body
|
||||
|
||||
|
||||
def json_request(conn, base_url, url, method, **kwargs):
|
||||
kwargs.setdefault('headers', {})
|
||||
kwargs['headers'].setdefault('Content-Type', 'application/json')
|
||||
kwargs['headers'].setdefault('Accept', 'application/json')
|
||||
|
||||
if 'body' in kwargs:
|
||||
kwargs['body'] = json.dumps(kwargs['body'])
|
||||
|
||||
resp, body = http_request(conn, base_url, url, method, **kwargs)
|
||||
content_type = resp.getheader('content-type', None)
|
||||
|
||||
if resp.status == 204 or resp.status == 205 or content_type is None:
|
||||
body = None
|
||||
elif 'application/json' in content_type:
|
||||
try:
|
||||
body = json.loads(body)
|
||||
except ValueError:
|
||||
LOG.error('Could not decode response body as JSON')
|
||||
else:
|
||||
body = None
|
||||
|
||||
return resp, body
|
||||
|
||||
|
||||
def create_resource_class(conn, base_url, name, service_type, racks, flavors):
|
||||
return json_request(conn, base_url, '/resource_classes', 'POST',
|
||||
body=dict(name=name, service_type=service_type,
|
||||
racks=racks, flavors=flavors))
|
||||
|
||||
|
||||
def create_rack(conn, base_url, name, slots, location,
|
||||
subnet, capacities, nodes):
|
||||
return json_request(conn, base_url, '/racks', 'POST',
|
||||
body=dict(name=name, slots=slots,
|
||||
subnet=subnet, location=location,
|
||||
capacities=capacities,nodes=nodes))
|
||||
|
||||
|
||||
def set_nodes_on_rack(conn, base_url, rack_url, nodes):
|
||||
return json_request(conn, base_url, rack_url, 'PUT',
|
||||
body=dict(nodes=nodes))
|
||||
|
||||
|
||||
def create_flavor(conn, base_url, resource_class_url, flavor):
|
||||
return json_request(conn, base_url + resource_class_url, '/flavors', 'PUT', body=flavor)
|
||||
|
||||
|
||||
def get_location(base_url, resp):
|
||||
return urlparse.urlparse(resp.getheader('location')).path[len(base_url):]
|
||||
|
||||
|
||||
def generate_uuid():
|
||||
return str(uuid.uuid4())
|
||||
|
||||
|
||||
def generate_data():
|
||||
logging.basicConfig(format='%(message)s', level=logging.DEBUG)
|
||||
|
||||
conn = httplib.HTTPConnection('localhost', 8585)
|
||||
|
||||
base_url = '/v1'
|
||||
|
||||
capacities = [
|
||||
dict(name='total_cpu', value='64', unit='count'),
|
||||
dict(name='total_memory', value='16384', unit='MB'),
|
||||
]
|
||||
nodes = []
|
||||
while len(nodes) < 30:
|
||||
nodes.append(dict(id=generate_uuid()))
|
||||
|
||||
rack_resp1, rack_body1 = create_rack(conn, base_url,
|
||||
name='compute_1', slots=3,
|
||||
subnet='192.168.1.0/24',
|
||||
location='room d2, row 1',
|
||||
capacities=capacities,
|
||||
nodes=nodes[0:3])
|
||||
|
||||
rack_resp2, rack_body2 = create_rack(conn, base_url,
|
||||
name='compute_2', slots=3,
|
||||
subnet='192.168.2.0/24',
|
||||
location='room d2, row 2',
|
||||
capacities=capacities,
|
||||
nodes=nodes[3:6])
|
||||
|
||||
compute_racks = [
|
||||
dict(id=rack_body1.get('id'),links=rack_body1.get('links')),
|
||||
dict(id=rack_body2.get('id'),links=rack_body2.get('links')),
|
||||
]
|
||||
|
||||
rack_resp3, rack_body3 = create_rack(conn, base_url,
|
||||
name='not_compute', slots=3,
|
||||
subnet='192.168.3.0/24',
|
||||
location='room d2, row 3',
|
||||
capacities=capacities,
|
||||
nodes=[nodes[7]])
|
||||
|
||||
non_compute_racks = [
|
||||
dict(id=rack_body3.get('id'),links=rack_body3.get('links')),
|
||||
]
|
||||
|
||||
|
||||
flavors = [
|
||||
dict(name='m1.small',
|
||||
capacities=[
|
||||
dict(name='cpu', value='1', unit='count'),
|
||||
dict(name='memory', value='1.7', unit='GB'),
|
||||
dict(name='storage', value='160', unit='GB'),
|
||||
]),
|
||||
dict(name='m1.medium',
|
||||
capacities=[
|
||||
dict(name='cpu', value='1', unit='count'),
|
||||
dict(name='memory', value='3.75', unit='GB'),
|
||||
dict(name='storage', value='410', unit='GB'),
|
||||
]),
|
||||
dict(name='m1.large',
|
||||
capacities=[
|
||||
dict(name='cpu', value='2', unit='count'),
|
||||
dict(name='memory', value='7.5', unit='GB'),
|
||||
dict(name='storage', value='840', unit='GB'),
|
||||
]),
|
||||
dict(name='m1.xlarge',
|
||||
capacities=[
|
||||
dict(name='cpu', value='4', unit='count'),
|
||||
dict(name='memory', value='15', unit='GB'),
|
||||
dict(name='storage', value='1680', unit='GB'),
|
||||
]),
|
||||
]
|
||||
|
||||
rc_resp1, rc_body1 = create_resource_class(conn, base_url,
|
||||
name='compute-rc',
|
||||
service_type='compute',
|
||||
racks=compute_racks,
|
||||
flavors=[flavors[3]])
|
||||
|
||||
rc_resp1, rc_body1 = create_resource_class(conn, base_url,
|
||||
name='non-compute-rc',
|
||||
service_type='not_compute',
|
||||
racks=non_compute_racks,
|
||||
flavors=[])
|
||||
|
||||
|
||||
if __name__ == '__main__':
|
||||
generate_data()
|
Loading…
x
Reference in New Issue
Block a user