refactor how datasets are stored

This commit is contained in:
Joshua Hesketh 2013-09-06 15:52:49 +10:00
parent e58ecb27ab
commit a74f49da7b
8 changed files with 51 additions and 40 deletions

View File

@ -1,7 +1,6 @@
{ {
"project": "openstack/nova", "project": "openstack/nova",
"type": "mysql", "type": "mysql",
"gate": "build:gate-real-db-upgrade_nova_mysql",
"db_user": "nova", "db_user": "nova",
"db_pass": "tester", "db_pass": "tester",
"database": "nova", "database": "nova",

View File

@ -13,6 +13,9 @@ Contents:
intro intro
structure structure
installation
configuration
running
Indices and tables Indices and tables

View File

@ -0,0 +1,9 @@
Installation
============
1. Download
2. Copy config
3. Edit config
4. Make dirs
5. Chmod them
6. Start using daemon

View File

@ -8,9 +8,14 @@
"jobs_working_dir": "/var/lib/turbo-hipster/jobs", "jobs_working_dir": "/var/lib/turbo-hipster/jobs",
"git_working_dir": "/var/lib/turbo-hipster/git", "git_working_dir": "/var/lib/turbo-hipster/git",
"pip_download_cache": "/var/cache/pip", "pip_download_cache": "/var/cache/pip",
"plugins": ["gate_real_db_upgrade"], "plugins": [
"publish_logs": {
{ "name": "gate_real_db_upgrade",
"datasets_dir": "/var/lib/turbo-hipster/datasets",
"gate": "build:gate-real-db-upgrade_nova_mysql",
}
],
"publish_logs": {
"type": "local", "type": "local",
"path": "/var/www/results/", "path": "/var/www/results/",
"prepend_url": "http://localhost/results/" "prepend_url": "http://localhost/results/"

View File

@ -25,10 +25,6 @@ packages =
[pbr] [pbr]
warnerrors = True warnerrors = True
[global]
setup-hooks =
pbr.hooks.setup_hook
[entry_points] [entry_points]
console_scripts = console_scripts =
turbo-hipster = turbo_hipster.worker_server:main turbo-hipster = turbo_hipster.worker_server:main

View File

@ -40,10 +40,11 @@ class Runner(threading.Thread):
log = logging.getLogger("task_plugins.gate_real_db_upgrade.task.Runner") log = logging.getLogger("task_plugins.gate_real_db_upgrade.task.Runner")
def __init__(self, config): def __init__(self, global_config, plugin_config):
super(Runner, self).__init__() super(Runner, self).__init__()
self._stop = threading.Event() self._stop = threading.Event()
self.config = config self.global_config = global_config
self.plugin_config = plugin_config
# Set up the runner worker # Set up the runner worker
self.gearman_worker = None self.gearman_worker = None
@ -65,16 +66,13 @@ class Runner(threading.Thread):
self.log.debug("Set up real_db gearman worker") self.log.debug("Set up real_db gearman worker")
self.gearman_worker = gear.Worker(__worker_name__) self.gearman_worker = gear.Worker(__worker_name__)
self.gearman_worker.addServer( self.gearman_worker.addServer(
self.config['zuul_server']['gearman_host'], self.global_config['zuul_server']['gearman_host'],
self.config['zuul_server']['gearman_port'] self.global_config['zuul_server']['gearman_port']
) )
self.register_functions() self.register_functions()
def register_functions(self): def register_functions(self):
""" Determine which functions to register based off available self.gearman_worker.registerFunction(self.plugin_config['gate'])
datasets """
for dataset in self._get_datasets():
self.gearman_worker.registerFunction(dataset['config']['gate'])
def stop(self): def stop(self):
self._stop.set() self._stop.set()
@ -160,7 +158,7 @@ class Runner(threading.Thread):
index_url = handle_results.generate_push_results( index_url = handle_results.generate_push_results(
self.job_datasets, self.job_datasets,
self.job.unique, self.job.unique,
self.config['publish_logs'] self.global_config['publish_logs']
) )
self.log.debug("Index URL found at %s" % index_url) self.log.debug("Index URL found at %s" % index_url)
self.work_data['url'] = index_url self.work_data['url'] = index_url
@ -186,8 +184,7 @@ class Runner(threading.Thread):
if len(self.datasets) > 0: if len(self.datasets) > 0:
return self.datasets return self.datasets
datasets_path = os.path.join(os.path.dirname(__file__), datasets_path = self.global_config['datasets_dir']
'datasets')
for ent in os.listdir(datasets_path): for ent in os.listdir(datasets_path):
dataset_dir = os.path.join(datasets_path, ent) dataset_dir = os.path.join(datasets_path, ent)
if (os.path.isdir(dataset_dir) and os.path.isfile( if (os.path.isdir(dataset_dir) and os.path.isfile(
@ -217,7 +214,7 @@ class Runner(threading.Thread):
dataset['config']['project'] and dataset['config']['project'] and
self._get_project_command(dataset['config']['type'])): self._get_project_command(dataset['config']['type'])):
dataset['log_file_path'] = os.path.join( dataset['log_file_path'] = os.path.join(
self.config['jobs_working_dir'], self.global_config['jobs_working_dir'],
self.job.unique, self.job.unique,
dataset['name'] + '.log' dataset['name'] + '.log'
) )
@ -262,7 +259,7 @@ class Runner(threading.Thread):
% { % {
'unique_id': self.job.unique, 'unique_id': self.job.unique,
'job_working_dir': os.path.join( 'job_working_dir': os.path.join(
self.config['jobs_working_dir'], self.global_config['jobs_working_dir'],
self.job.unique self.job.unique
), ),
'git_path': git_path, 'git_path': git_path,
@ -277,7 +274,7 @@ class Runner(threading.Thread):
dataset['dataset_dir'], dataset['dataset_dir'],
dataset['config']['logging_conf'] dataset['config']['logging_conf']
), ),
'pip_cache_dir': self.config['pip_download_cache'] 'pip_cache_dir': self.global_config['pip_download_cache']
} }
) )
@ -285,13 +282,13 @@ class Runner(threading.Thread):
syslog = '/var/log/syslog' syslog = '/var/log/syslog'
sqlslo = '/var/log/mysql/slow-queries.log' sqlslo = '/var/log/mysql/slow-queries.log'
sqlerr = '/var/log/mysql/error.log' sqlerr = '/var/log/mysql/error.log'
if 'logs' in self.config: if 'logs' in self.global_config:
if 'syslog' in self.config['logs']: if 'syslog' in self.global_config['logs']:
syslog = self.config['logs']['syslog'] syslog = self.global_config['logs']['syslog']
if 'sqlslo' in self.config['logs']: if 'sqlslo' in self.global_config['logs']:
sqlslo = self.config['logs']['sqlslo'] sqlslo = self.global_config['logs']['sqlslo']
if 'sqlerr' in self.config['logs']: if 'sqlerr' in self.global_config['logs']:
sqlerr = self.config['logs']['sqlerr'] sqlerr = self.global_config['logs']['sqlerr']
utils.execute_to_log( utils.execute_to_log(
cmd, cmd,
@ -309,9 +306,9 @@ class Runner(threading.Thread):
self.log.debug("Grab the patchset we want to test against") self.log.debug("Grab the patchset we want to test against")
repo = utils.GitRepository( repo = utils.GitRepository(
self.config['zuul_server']['git_url'] + project_name + '/.git', self.global_config['zuul_server']['git_url'] + project_name + '/.git',
os.path.join( os.path.join(
self.config['git_working_dir'], self.global_config['git_working_dir'],
__worker_name__, __worker_name__,
project_name project_name
) )

View File

@ -67,19 +67,21 @@ class Server(object):
""" Load the available plugins from task_plugins """ """ Load the available plugins from task_plugins """
# Load plugins # Load plugins
for plugin in self.config['plugins']: for plugin in self.config['plugins']:
print self.plugins.append({
plugin_info = imp.find_module('task', 'module': __import__('turbo_hipster.task_plugins.' +
[(os.path.dirname( plugin['name'],
os.path.realpath(__file__)) + fromlist='turbo_hipster.task_plugins'),
'/task_plugins/' + plugin)]) 'plugin_config': self.plugin
self.plugins.append(imp.load_module('task', *plugin_info)) })
def run_tasks(self): def run_tasks(self):
""" Run the tasks """ """ Run the tasks """
for plugin in self.plugins: for plugin in self.plugins:
self.tasks[plugin.__worker_name__] = plugin.Runner(self.config) module = plugin['module']
self.tasks[plugin.__worker_name__].daemon = True self.tasks[module.__worker_name__] = module.Runner(self.config,
self.tasks[plugin.__worker_name__].start() plugin['plugin_config'])
self.tasks[module.__worker_name__].daemon = True
self.tasks[module.__worker_name__].start()
self.manager = worker_manager.GearmanManager(self.config, self.tasks) self.manager = worker_manager.GearmanManager(self.config, self.tasks)
self.manager.daemon = True self.manager.daemon = True