fb61e864b9
The _destroy_evacuated_instances method on compute startup tries to cleanup guests on the hypervisor and allocations held against that compute node resource provider by evacuated instances, but doesn't take into account that those evacuated instances could have been deleted in the meantime which leads to a lazy-load InstanceNotFound error that kills the startup of the compute service. This change does two things in the _destroy_evacuated_instances method: 1. Loads the evacuated instances with a read_deleted='yes' context when calling _get_instances_on_driver(). This should be fine since _get_instances_on_driver() is already returning deleted instances anyway (InstanceList.get_by_filters defaults to read deleted instances unless the filters tell it otherwise - which we don't in this case). This is needed so that things like driver.destroy() don't raise InstanceNotFound while lazy-loading fields on the instance. 2. Skips the call to remove_allocation_from_compute() if the evacuated instance is already deleted. If the instance is already deleted, its allocations should have been cleaned up by its hosting compute service (or the API). The functional regression test is updated to show the bug is now fixed. Conflicts: nova/compute/manager.py NOTE(mriedem): The conflict is due to not having change I1073faca6760bff3da0aaf3e8357bd8e64854be3 in Pike. Change-Id: I1f4b3540dd453650f94333b36d7504ba164192f7 Closes-Bug: #1794996 (cherry picked from commit |
||
---|---|---|
.. | ||
monitors | ||
__init__.py | ||
api.py | ||
build_results.py | ||
cells_api.py | ||
claims.py | ||
flavors.py | ||
instance_actions.py | ||
manager.py | ||
power_state.py | ||
provider_tree.py | ||
resource_tracker.py | ||
rpcapi.py | ||
stats.py | ||
task_states.py | ||
utils.py | ||
vm_states.py |