7b58c1c9ea
Partially-addresses: BP wider-aggregate-range Support for two new aggregation functions has been added: * stddev: population standard deviation (unparameterized) * cardinality: count of distinct values for a sample attribute (parameterized by attribute name) For example in order to calculate the standard deviation of CPU util: GET /v2/meters/cpu_util/statistics?aggregate.func=stddev HTTP/1.0 200 OK [{"aggregate": {"stddev": 0.6858829535841072}, "duration_start": "2014-01-30T11:13:23", "duration_end": "2014-01-31T16:07:13", "duration": 104030.0, "period": 0, "period_start": "2014-01-30T11:13:23", "period_end": "2014-01-31T16:07:13", "groupby": null, "unit": "%"}] Or to calculate the number of distinct instances that existed for each tenant contrasted with the total number of instance samples for that tenant in each 15 min period: GET /v2/meters/instance/statistics?aggregate.func=cardinality&aggregate.param=resource_id &aggregate.func=count &groupby=project_id&period=900 HTTP/1.0 200 OK [{"count": 19, "aggregate": {"count": 19.0, "cardinality/resource_id": 3.0}, "duration": 328.478029, "duration_start": "2014-01-31T10:00:41.823919", "duration_end": "2014-01-31T10:06:10.301948", "period": 900, "period_start": "2014-01-31T10:00:00", "period_end": "2014-01-31T10:15:00", "groupby": {"project_id": "061a5c91811e4044b7dc86c6136c4f99"}, "unit": "instance"}, {"count": 22, "aggregate": {"count": 22.0, "cardinality/resource_id": 4.0}, "duration": 808.00384, "duration_start": "2014-01-31T10:15:15", "duration_end": "2014-01-31T10:28:43.003840", "period": 900, "period_start": "2014-01-31T10:15:00", "period_end": "2014-01-31T10:30:00", "groupby": {"project_id": "061a5c91811e4044b7dc86c6136c4f99"}, "unit": "instance"}, {"count": 2, "aggregate": {"count": 2.0, "cardinality/resource_id": 2.0}, "duration": 0.0, "duration_start": "2014-01-31T10:35:15", "duration_end": "2014-01-31T10:35:15", "period": 900, "period_start": "2014-01-31T10:30:00", "period_end": "2014-01-31T10:45:00", "groupby": {"project_id": "061a5c91811e4044b7dc86c6136c4f99"}, "unit": "instance"}] Test coverage is provided by the scenario test added in next patch of this series. Change-Id: Id3a37622f35d9a7d757c485b4d486b3f01cc6474
176 lines
5.9 KiB
Python
176 lines
5.9 KiB
Python
# -*- encoding: utf-8 -*-
|
|
#
|
|
# Copyright © 2012 New Dream Network, LLC (DreamHost)
|
|
#
|
|
# Author: Doug Hellmann <doug.hellmann@dreamhost.com>
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
|
# not use this file except in compliance with the License. You may obtain
|
|
# a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
|
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
|
# License for the specific language governing permissions and limitations
|
|
# under the License.
|
|
"""Storage backend management
|
|
"""
|
|
|
|
import six.moves.urllib.parse as urlparse
|
|
|
|
from oslo.config import cfg
|
|
import six
|
|
from stevedore import driver
|
|
|
|
from ceilometer.openstack.common.gettextutils import _ # noqa
|
|
from ceilometer.openstack.common import log
|
|
from ceilometer import service
|
|
from ceilometer import utils
|
|
|
|
|
|
LOG = log.getLogger(__name__)
|
|
|
|
STORAGE_ENGINE_NAMESPACE = 'ceilometer.storage'
|
|
|
|
OLD_STORAGE_OPTS = [
|
|
cfg.StrOpt('database_connection',
|
|
secret=True,
|
|
default=None,
|
|
help='DEPRECATED - Database connection string.',
|
|
),
|
|
]
|
|
|
|
cfg.CONF.register_opts(OLD_STORAGE_OPTS)
|
|
|
|
|
|
STORAGE_OPTS = [
|
|
cfg.IntOpt('time_to_live',
|
|
default=-1,
|
|
help="Number of seconds that samples are kept "
|
|
"in the database for (<= 0 means forever)."),
|
|
]
|
|
|
|
cfg.CONF.register_opts(STORAGE_OPTS, group='database')
|
|
|
|
cfg.CONF.import_opt('connection',
|
|
'ceilometer.openstack.common.db.sqlalchemy.session',
|
|
group='database')
|
|
|
|
|
|
class StorageBadVersion(Exception):
|
|
"""Error raised when the storage backend version is not good enough."""
|
|
|
|
|
|
class StorageBadAggregate(Exception):
|
|
"""Error raised when an aggregate is unacceptable to storage backend."""
|
|
code = 400
|
|
|
|
|
|
def get_engine(conf):
|
|
"""Load the configured engine and return an instance."""
|
|
if conf.database_connection:
|
|
conf.set_override('connection', conf.database_connection,
|
|
group='database')
|
|
engine_name = urlparse.urlparse(conf.database.connection).scheme
|
|
LOG.debug(_('looking for %(name)r driver in %(namespace)r') % (
|
|
{'name': engine_name,
|
|
'namespace': STORAGE_ENGINE_NAMESPACE}))
|
|
mgr = driver.DriverManager(STORAGE_ENGINE_NAMESPACE,
|
|
engine_name,
|
|
invoke_on_load=True)
|
|
return mgr.driver
|
|
|
|
|
|
def get_connection(conf):
|
|
"""Return an open connection to the database."""
|
|
return get_engine(conf).get_connection(conf)
|
|
|
|
|
|
class SampleFilter(object):
|
|
"""Holds the properties for building a query from a meter/sample filter.
|
|
|
|
:param user: The sample owner.
|
|
:param project: The sample project.
|
|
:param start: Earliest time point in the request.
|
|
:param start_timestamp_op: Earliest timestamp operation in the request.
|
|
:param end: Latest time point in the request.
|
|
:param end_timestamp_op: Latest timestamp operation in the request.
|
|
:param resource: Optional filter for resource id.
|
|
:param meter: Optional filter for meter type using the meter name.
|
|
:param source: Optional source filter.
|
|
:param message_id: Optional sample_id filter.
|
|
:param metaquery: Optional filter on the metadata
|
|
"""
|
|
def __init__(self, user=None, project=None,
|
|
start=None, start_timestamp_op=None,
|
|
end=None, end_timestamp_op=None,
|
|
resource=None, meter=None,
|
|
source=None, message_id=None,
|
|
metaquery={}):
|
|
self.user = user
|
|
self.project = project
|
|
self.start = utils.sanitize_timestamp(start)
|
|
self.start_timestamp_op = start_timestamp_op
|
|
self.end = utils.sanitize_timestamp(end)
|
|
self.end_timestamp_op = end_timestamp_op
|
|
self.resource = resource
|
|
self.meter = meter
|
|
self.source = source
|
|
self.metaquery = metaquery
|
|
self.message_id = message_id
|
|
|
|
|
|
class EventFilter(object):
|
|
"""Properties for building an Event query.
|
|
|
|
:param start_time: UTC start datetime (mandatory)
|
|
:param end_time: UTC end datetime (mandatory)
|
|
:param event_type: the name of the event. None for all.
|
|
:param message_id: the message_id of the event. None for all.
|
|
:param traits_filter: the trait filter dicts, all of which are optional.
|
|
This parameter is a list of dictionaries that specify
|
|
trait values:
|
|
{'key': <key>,
|
|
'string': <value>,
|
|
'integer': <value>,
|
|
'datetime': <value>,
|
|
'float': <value>,
|
|
'op': <eq, lt, le, ne, gt or ge> }
|
|
"""
|
|
|
|
def __init__(self, start_time=None, end_time=None, event_type=None,
|
|
message_id=None, traits_filter=[]):
|
|
self.start_time = utils.sanitize_timestamp(start_time)
|
|
self.end_time = utils.sanitize_timestamp(end_time)
|
|
self.message_id = message_id
|
|
self.event_type = event_type
|
|
self.traits_filter = traits_filter
|
|
|
|
def __repr__(self):
|
|
return ("<EventFilter(start_time: %s,"
|
|
" end_time: %s,"
|
|
" event_type: %s,"
|
|
" traits: %s)>" %
|
|
(self.start_time,
|
|
self.end_time,
|
|
self.event_type,
|
|
six.text_type(self.traits_filter)))
|
|
|
|
|
|
def dbsync():
|
|
service.prepare_service()
|
|
get_connection(cfg.CONF).upgrade()
|
|
|
|
|
|
def expirer():
|
|
service.prepare_service()
|
|
if cfg.CONF.database.time_to_live > 0:
|
|
LOG.debug(_("Clearing expired metering data"))
|
|
storage_conn = get_connection(cfg.CONF)
|
|
storage_conn.clear_expired_metering_data(
|
|
cfg.CONF.database.time_to_live)
|
|
else:
|
|
LOG.info(_("Nothing to clean, database time to live is disabled"))
|