Revert "Revert "Remove ClusteredComputeManager""
We're now past the 3 month deprecation period and can remove this.
This reverts commit c5ab2fdd43
.
Change-Id: I4f57f322027b3db56272e2a2e85f972c37d7b9fd
This commit is contained in:
parent
eae8e07132
commit
4c61f1ac18
@ -1,122 +0,0 @@
|
|||||||
# coding=utf-8
|
|
||||||
#
|
|
||||||
# Copyright 2014 Red Hat, Inc.
|
|
||||||
# Copyright 2013 Hewlett-Packard Development Company, L.P.
|
|
||||||
# All Rights Reserved.
|
|
||||||
#
|
|
||||||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
|
||||||
# not use this file except in compliance with the License. You may obtain
|
|
||||||
# a copy of the License at
|
|
||||||
#
|
|
||||||
# http://www.apache.org/licenses/LICENSE-2.0
|
|
||||||
#
|
|
||||||
# Unless required by applicable law or agreed to in writing, software
|
|
||||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
|
||||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
|
||||||
# License for the specific language governing permissions and limitations
|
|
||||||
# under the License.
|
|
||||||
|
|
||||||
"""Short term workaround for friction in the Nova compute manager with Ironic.
|
|
||||||
|
|
||||||
https://etherpad.openstack.org/p/ironic-nova-friction contains current design
|
|
||||||
work. The goal here is to generalise the areas where n-c talking to a clustered
|
|
||||||
hypervisor has issues, and long term fold them into the main ComputeManager.
|
|
||||||
"""
|
|
||||||
|
|
||||||
from oslo_concurrency import lockutils
|
|
||||||
from oslo_log import log as logging
|
|
||||||
|
|
||||||
from nova.compute import manager
|
|
||||||
import nova.context
|
|
||||||
from nova.i18n import _LW
|
|
||||||
|
|
||||||
|
|
||||||
LOG = logging.getLogger(__name__)
|
|
||||||
|
|
||||||
CCM_SEMAPHORE = 'clustered_compute_manager'
|
|
||||||
|
|
||||||
|
|
||||||
# TODO(jroll) delete this in Ocata
|
|
||||||
class ClusteredComputeManager(manager.ComputeManager):
|
|
||||||
|
|
||||||
def __init__(self, *args, **kwargs):
|
|
||||||
LOG.warning(_LW('ClusteredComputeManager is deprecated. As of the '
|
|
||||||
'Newton release of Nova, the ironic '
|
|
||||||
'virt driver directly supports using multiple '
|
|
||||||
'compute hosts without ClusteredComputeManager. '
|
|
||||||
'Users should unset the `compute_manager` '
|
|
||||||
'configuration option to use Nova\'s default '
|
|
||||||
'instead. ClusteredComputeManager will be removed '
|
|
||||||
'during the Ocata cycle.'))
|
|
||||||
super(ClusteredComputeManager, self).__init__(*args, **kwargs)
|
|
||||||
|
|
||||||
def init_host(self):
|
|
||||||
"""Initialization for a clustered compute service."""
|
|
||||||
self.driver.init_host(host=self.host)
|
|
||||||
# Not used currently.
|
|
||||||
# context = nova.context.get_admin_context()
|
|
||||||
# instances = instance_obj.InstanceList.get_by_host(
|
|
||||||
# context, self.host, expected_attrs=['info_cache'])
|
|
||||||
|
|
||||||
# defer_iptables_apply is moot for clusters - no local iptables
|
|
||||||
# if CONF.defer_iptables_apply:
|
|
||||||
# self.driver.filter_defer_apply_on()
|
|
||||||
|
|
||||||
self.init_virt_events()
|
|
||||||
|
|
||||||
# try:
|
|
||||||
# evacuation is moot for a clustered hypervisor
|
|
||||||
# # checking that instance was not already evacuated to other host
|
|
||||||
# self._destroy_evacuated_instances(context)
|
|
||||||
# Don't run _init_instance until we solve the partitioning problem
|
|
||||||
# - with N n-cpu's all claiming the same hostname, running
|
|
||||||
# _init_instance here would lead to race conditions where each runs
|
|
||||||
# _init_instance concurrently.
|
|
||||||
# for instance in instances:
|
|
||||||
# self._init_instance(context, instance)
|
|
||||||
# finally:
|
|
||||||
# defer_iptables_apply is moot for clusters - no local iptables
|
|
||||||
# if CONF.defer_iptables_apply:
|
|
||||||
# self.driver.filter_defer_apply_off()
|
|
||||||
|
|
||||||
def pre_start_hook(self):
|
|
||||||
"""Update our available resources
|
|
||||||
|
|
||||||
After the service is initialized, but before we fully bring
|
|
||||||
the service up by listening on RPC queues, make sure to update
|
|
||||||
our available resources (and indirectly our available nodes).
|
|
||||||
"""
|
|
||||||
# This is an optimisation to immediately advertise resources but
|
|
||||||
# the periodic task will update them eventually anyway, so ignore
|
|
||||||
# errors as they may be transient (e.g. the scheduler isn't
|
|
||||||
# available...). XXX(lifeless) this applies to all ComputeManagers
|
|
||||||
# and once I feature freeze is over we should push that to nova
|
|
||||||
# directly.
|
|
||||||
try:
|
|
||||||
self.update_available_resource(nova.context.get_admin_context())
|
|
||||||
except Exception:
|
|
||||||
pass
|
|
||||||
|
|
||||||
@lockutils.synchronized(CCM_SEMAPHORE)
|
|
||||||
def _update_resources(self, node_uuid):
|
|
||||||
"""Update the specified node resource
|
|
||||||
|
|
||||||
Updates the resources for instance while protecting against a race on
|
|
||||||
self._resource_tracker_dict.
|
|
||||||
:param node_uuid: UUID of the Ironic node to update resources for.
|
|
||||||
"""
|
|
||||||
self.update_available_resource_for_node(
|
|
||||||
nova.context.get_admin_context(), node_uuid)
|
|
||||||
|
|
||||||
def terminate_instance(self, context, instance, bdms, reservations):
|
|
||||||
"""Terminate an instance on a node.
|
|
||||||
|
|
||||||
We override this method and force a post-termination update to Nova's
|
|
||||||
resources. This avoids having to wait for a Nova periodic task tick
|
|
||||||
before nodes can be reused.
|
|
||||||
"""
|
|
||||||
super(ClusteredComputeManager, self).terminate_instance(context,
|
|
||||||
instance,
|
|
||||||
bdms,
|
|
||||||
reservations)
|
|
||||||
self._update_resources(instance.node)
|
|
@ -0,0 +1,5 @@
|
|||||||
|
---
|
||||||
|
upgrade:
|
||||||
|
- |
|
||||||
|
The deprecated ironic.nova.ClusteredComputerManager module is now removed.
|
||||||
|
This is not required with nova >= 14.0.0 (Newton).
|
Loading…
Reference in New Issue
Block a user