![Eoghan Glynn](/assets/img/avatar_default.png)
Addresses: BP decoupled-source-sink-discoverable-resources Add support for a new decoupled model of pipeline config which does not conflate the elements specific to sources and sinks. Instead of the basic unit of config being a consolidated pipeline, the sources and sinks may now be specified as separate lists of dictionaries, before being linked together once parsed. This allows source-specific configuration, such as resource discovery, to be kept focused only on the fine-grained source while avoiding the necessity for wide duplication of sink- related config. The shape of the new config format is best illustrated with an example: --- sources: - name: meter_source interval: 600 meters: - "*" sinks: - meter_sink - name: host_cpu_source interval: 120 meters: - "cpu.util.*min" resources: - "snmp://ip1" - "snmp://ip2" - "snmp://ip3" sinks: - meter_sink - lossy_sink - name: instance_cpu_source interval: 60 meters: - "cpu" sinks: - cpu_sink sinks: - name: meter_sink transformers: publishers: - rpc:// - name: lossy_sink transformers: publishers: - udp://addr - name: cpu_sink transformers: - name: "rate_of_change" parameters: target: name: "cpu_util" unit: "%" type: "gauge" scale: "100.0 / (10**9 * (resource_metadata.cpu_number or 1))" publishers: - rpc:// The old pipeline.yaml format continues to be supported as a deprecated feature to avoid breaking existing deployments. The semantics of the common individual configuration elements are identical in the deprecated and decoupled versions. Change-Id: Ide86c0feba88ae736f2a913b5faa95e640c4ceaf
152 lines
5.5 KiB
Python
152 lines
5.5 KiB
Python
# -*- encoding: utf-8 -*-
|
|
#
|
|
# Copyright © 2013 Julien Danjou
|
|
# Copyright © 2014 Red Hat, Inc
|
|
#
|
|
# Authors: Julien Danjou <julien@danjou.info>
|
|
# Eoghan Glynn <eglynn@redhat.com>
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
|
# not use this file except in compliance with the License. You may obtain
|
|
# a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
|
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
|
# License for the specific language governing permissions and limitations
|
|
# under the License.
|
|
|
|
import collections
|
|
import itertools
|
|
import urlparse
|
|
|
|
from stevedore import extension
|
|
|
|
from ceilometer.openstack.common import context
|
|
from ceilometer.openstack.common.gettextutils import _ # noqa
|
|
from ceilometer.openstack.common import log
|
|
from ceilometer.openstack.common import service as os_service
|
|
from ceilometer import pipeline
|
|
|
|
LOG = log.getLogger(__name__)
|
|
|
|
|
|
class PollingTask(object):
|
|
"""Polling task for polling samples and inject into pipeline.
|
|
A polling task can be invoked periodically or only once.
|
|
"""
|
|
|
|
def __init__(self, agent_manager):
|
|
self.manager = agent_manager
|
|
self.pollsters = set()
|
|
# Resource definitions are indexed by the pollster
|
|
# Use dict of set here to remove the duplicated resource definitions
|
|
# for each pollster.
|
|
self.resources = collections.defaultdict(set)
|
|
self.publish_context = pipeline.PublishContext(
|
|
agent_manager.context)
|
|
|
|
def add(self, pollster, pipelines):
|
|
self.publish_context.add_pipelines(pipelines)
|
|
for pipeline in pipelines:
|
|
self.resources[pollster.name].update(pipeline.resources)
|
|
self.pollsters.update([pollster])
|
|
|
|
def poll_and_publish(self):
|
|
"""Polling sample and publish into pipeline."""
|
|
agent_resources = self.manager.discover()
|
|
with self.publish_context as publisher:
|
|
cache = {}
|
|
for pollster in self.pollsters:
|
|
LOG.info(_("Polling pollster %s"), pollster.name)
|
|
source_resources = list(self.resources[pollster.name])
|
|
try:
|
|
samples = list(pollster.obj.get_samples(
|
|
self.manager,
|
|
cache,
|
|
resources=source_resources or agent_resources,
|
|
))
|
|
publisher(samples)
|
|
except Exception as err:
|
|
LOG.warning(_(
|
|
'Continue after error from %(name)s: %(error)s')
|
|
% ({'name': pollster.name, 'error': err}),
|
|
exc_info=True)
|
|
|
|
|
|
class AgentManager(os_service.Service):
|
|
|
|
def __init__(self, namespace, default_discovery=[]):
|
|
super(AgentManager, self).__init__()
|
|
self.default_discovery = default_discovery
|
|
self.pollster_manager = self._extensions('poll', namespace)
|
|
self.discovery_manager = self._extensions('discover')
|
|
self.context = context.RequestContext('admin', 'admin', is_admin=True)
|
|
|
|
@staticmethod
|
|
def _extensions(category, agent_ns=None):
|
|
namespace = ('ceilometer.%s.%s' % (category, agent_ns) if agent_ns
|
|
else 'ceilometer.%s' % category)
|
|
return extension.ExtensionManager(
|
|
namespace=namespace,
|
|
invoke_on_load=True,
|
|
)
|
|
|
|
def create_polling_task(self):
|
|
"""Create an initially empty polling task."""
|
|
return PollingTask(self)
|
|
|
|
def setup_polling_tasks(self):
|
|
polling_tasks = {}
|
|
for pipeline, pollster in itertools.product(
|
|
self.pipeline_manager.pipelines,
|
|
self.pollster_manager.extensions):
|
|
if pipeline.support_meter(pollster.name):
|
|
polling_task = polling_tasks.get(pipeline.get_interval())
|
|
if not polling_task:
|
|
polling_task = self.create_polling_task()
|
|
polling_tasks[pipeline.get_interval()] = polling_task
|
|
polling_task.add(pollster, [pipeline])
|
|
|
|
return polling_tasks
|
|
|
|
def start(self):
|
|
self.pipeline_manager = pipeline.setup_pipeline()
|
|
|
|
for interval, task in self.setup_polling_tasks().iteritems():
|
|
self.tg.add_timer(interval,
|
|
self.interval_task,
|
|
task=task)
|
|
|
|
@staticmethod
|
|
def interval_task(task):
|
|
task.poll_and_publish()
|
|
|
|
@staticmethod
|
|
def _parse_discoverer(url):
|
|
s = urlparse.urlparse(url)
|
|
return (s.scheme or s.path), (s.netloc + s.path if s.scheme else None)
|
|
|
|
def _discoverer(self, name):
|
|
for d in self.discovery_manager:
|
|
if d.name == name:
|
|
return d.obj
|
|
return None
|
|
|
|
def discover(self, discovery=[]):
|
|
resources = []
|
|
for url in (discovery or self.default_discovery):
|
|
name, param = self._parse_discoverer(url)
|
|
discoverer = self._discoverer(name)
|
|
if discoverer:
|
|
try:
|
|
discovered = discoverer.discover(param)
|
|
resources.extend(discovered)
|
|
except Exception as err:
|
|
LOG.exception(_('Unable to discover resources: %s') % err)
|
|
else:
|
|
LOG.warning(_('Unknown discovery extension: %s') % name)
|
|
return resources
|