Revert "Revert "Remove ClusteredComputeManager""

We're now past the 3 month deprecation period and can remove this.

This reverts commit c5ab2fdd43.

Change-Id: I4f57f322027b3db56272e2a2e85f972c37d7b9fd
This commit is contained in:
Jim Rollenhagen 2017-01-23 08:34:04 -05:00
parent eae8e07132
commit 4c61f1ac18
4 changed files with 5 additions and 122 deletions

View File

@ -1,122 +0,0 @@
# coding=utf-8
#
# Copyright 2014 Red Hat, Inc.
# Copyright 2013 Hewlett-Packard Development Company, L.P.
# All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
"""Short term workaround for friction in the Nova compute manager with Ironic.
https://etherpad.openstack.org/p/ironic-nova-friction contains current design
work. The goal here is to generalise the areas where n-c talking to a clustered
hypervisor has issues, and long term fold them into the main ComputeManager.
"""
from oslo_concurrency import lockutils
from oslo_log import log as logging
from nova.compute import manager
import nova.context
from nova.i18n import _LW
LOG = logging.getLogger(__name__)
CCM_SEMAPHORE = 'clustered_compute_manager'
# TODO(jroll) delete this in Ocata
class ClusteredComputeManager(manager.ComputeManager):
def __init__(self, *args, **kwargs):
LOG.warning(_LW('ClusteredComputeManager is deprecated. As of the '
'Newton release of Nova, the ironic '
'virt driver directly supports using multiple '
'compute hosts without ClusteredComputeManager. '
'Users should unset the `compute_manager` '
'configuration option to use Nova\'s default '
'instead. ClusteredComputeManager will be removed '
'during the Ocata cycle.'))
super(ClusteredComputeManager, self).__init__(*args, **kwargs)
def init_host(self):
"""Initialization for a clustered compute service."""
self.driver.init_host(host=self.host)
# Not used currently.
# context = nova.context.get_admin_context()
# instances = instance_obj.InstanceList.get_by_host(
# context, self.host, expected_attrs=['info_cache'])
# defer_iptables_apply is moot for clusters - no local iptables
# if CONF.defer_iptables_apply:
# self.driver.filter_defer_apply_on()
self.init_virt_events()
# try:
# evacuation is moot for a clustered hypervisor
# # checking that instance was not already evacuated to other host
# self._destroy_evacuated_instances(context)
# Don't run _init_instance until we solve the partitioning problem
# - with N n-cpu's all claiming the same hostname, running
# _init_instance here would lead to race conditions where each runs
# _init_instance concurrently.
# for instance in instances:
# self._init_instance(context, instance)
# finally:
# defer_iptables_apply is moot for clusters - no local iptables
# if CONF.defer_iptables_apply:
# self.driver.filter_defer_apply_off()
def pre_start_hook(self):
"""Update our available resources
After the service is initialized, but before we fully bring
the service up by listening on RPC queues, make sure to update
our available resources (and indirectly our available nodes).
"""
# This is an optimisation to immediately advertise resources but
# the periodic task will update them eventually anyway, so ignore
# errors as they may be transient (e.g. the scheduler isn't
# available...). XXX(lifeless) this applies to all ComputeManagers
# and once I feature freeze is over we should push that to nova
# directly.
try:
self.update_available_resource(nova.context.get_admin_context())
except Exception:
pass
@lockutils.synchronized(CCM_SEMAPHORE)
def _update_resources(self, node_uuid):
"""Update the specified node resource
Updates the resources for instance while protecting against a race on
self._resource_tracker_dict.
:param node_uuid: UUID of the Ironic node to update resources for.
"""
self.update_available_resource_for_node(
nova.context.get_admin_context(), node_uuid)
def terminate_instance(self, context, instance, bdms, reservations):
"""Terminate an instance on a node.
We override this method and force a post-termination update to Nova's
resources. This avoids having to wait for a Nova periodic task tick
before nodes can be reused.
"""
super(ClusteredComputeManager, self).terminate_instance(context,
instance,
bdms,
reservations)
self._update_resources(instance.node)

View File

@ -0,0 +1,5 @@
---
upgrade:
- |
The deprecated ironic.nova.ClusteredComputerManager module is now removed.
This is not required with nova >= 14.0.0 (Newton).