438 lines
21 KiB
Python
438 lines
21 KiB
Python
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
|
# not use this file except in compliance with the License. You may obtain
|
|
# a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
|
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
|
# License for the specific language governing permissions and limitations
|
|
# under the License.
|
|
|
|
import mock
|
|
from oslo_serialization import jsonutils
|
|
from oslo_utils.fixture import uuidsentinel as uuids
|
|
|
|
from nova.compute import rpcapi as compute_rpcapi
|
|
from nova.conductor.tasks import migrate
|
|
from nova import context
|
|
from nova import exception
|
|
from nova import objects
|
|
from nova.scheduler.client import query
|
|
from nova.scheduler.client import report
|
|
from nova.scheduler import utils as scheduler_utils
|
|
from nova import test
|
|
from nova.tests.unit.conductor.test_conductor import FakeContext
|
|
from nova.tests.unit import fake_flavor
|
|
from nova.tests.unit import fake_instance
|
|
|
|
|
|
class MigrationTaskTestCase(test.NoDBTestCase):
|
|
def setUp(self):
|
|
super(MigrationTaskTestCase, self).setUp()
|
|
self.user_id = 'fake'
|
|
self.project_id = 'fake'
|
|
self.context = FakeContext(self.user_id, self.project_id)
|
|
# Normally RequestContext.cell_uuid would be set when targeting
|
|
# the context in nova.conductor.manager.targets_cell but we just
|
|
# fake it here.
|
|
self.context.cell_uuid = uuids.cell1
|
|
self.flavor = fake_flavor.fake_flavor_obj(self.context)
|
|
self.flavor.extra_specs = {'extra_specs': 'fake'}
|
|
inst = fake_instance.fake_db_instance(
|
|
image_ref='image_ref', flavor=self.flavor)
|
|
inst_object = objects.Instance(
|
|
flavor=self.flavor,
|
|
numa_topology=None,
|
|
pci_requests=None,
|
|
system_metadata={'image_hw_disk_bus': 'scsi'})
|
|
self.instance = objects.Instance._from_db_object(
|
|
self.context, inst_object, inst, [])
|
|
self.request_spec = objects.RequestSpec(image=objects.ImageMeta())
|
|
self.host_lists = [[objects.Selection(service_host="host1",
|
|
nodename="node1", cell_uuid=uuids.cell1)]]
|
|
self.filter_properties = {'limits': {}, 'retry': {'num_attempts': 1,
|
|
'hosts': [['host1', 'node1']]}}
|
|
self.reservations = []
|
|
self.clean_shutdown = True
|
|
|
|
_p = mock.patch('nova.compute.utils.heal_reqspec_is_bfv')
|
|
self.heal_reqspec_is_bfv_mock = _p.start()
|
|
self.addCleanup(_p.stop)
|
|
|
|
_p = mock.patch('nova.objects.RequestSpec.ensure_network_information')
|
|
self.ensure_network_information_mock = _p.start()
|
|
self.addCleanup(_p.stop)
|
|
|
|
self.mock_network_api = mock.Mock()
|
|
|
|
def _generate_task(self):
|
|
return migrate.MigrationTask(self.context, self.instance, self.flavor,
|
|
self.request_spec,
|
|
self.clean_shutdown,
|
|
compute_rpcapi.ComputeAPI(),
|
|
query.SchedulerQueryClient(),
|
|
report.SchedulerReportClient(),
|
|
host_list=None,
|
|
network_api=self.mock_network_api)
|
|
|
|
@mock.patch.object(objects.MigrationList, 'get_by_filters')
|
|
@mock.patch('nova.scheduler.client.report.SchedulerReportClient')
|
|
@mock.patch('nova.objects.ComputeNode.get_by_host_and_nodename')
|
|
@mock.patch('nova.objects.Migration.save')
|
|
@mock.patch('nova.objects.Migration.create')
|
|
@mock.patch('nova.objects.Service.get_minimum_version_multi')
|
|
@mock.patch('nova.availability_zones.get_host_availability_zone')
|
|
@mock.patch.object(scheduler_utils, 'setup_instance_group')
|
|
@mock.patch.object(query.SchedulerQueryClient, 'select_destinations')
|
|
@mock.patch.object(compute_rpcapi.ComputeAPI, 'prep_resize')
|
|
@mock.patch('nova.conductor.tasks.cross_cell_migrate.'
|
|
'CrossCellMigrationTask.execute')
|
|
def _test_execute(self, cross_cell_exec_mock, prep_resize_mock,
|
|
sel_dest_mock, sig_mock, az_mock, gmv_mock, cm_mock,
|
|
sm_mock, cn_mock, rc_mock, gbf_mock,
|
|
requested_destination=False, same_cell=True):
|
|
sel_dest_mock.return_value = self.host_lists
|
|
az_mock.return_value = 'myaz'
|
|
gbf_mock.return_value = objects.MigrationList()
|
|
mock_get_resources = \
|
|
self.mock_network_api.get_requested_resource_for_instance
|
|
mock_get_resources.return_value = []
|
|
|
|
if requested_destination:
|
|
self.request_spec.requested_destination = objects.Destination(
|
|
host='target_host', node=None,
|
|
allow_cross_cell_move=not same_cell)
|
|
self.request_spec.retry = objects.SchedulerRetries.from_dict(
|
|
self.context, self.filter_properties['retry'])
|
|
self.filter_properties.pop('retry')
|
|
self.filter_properties['requested_destination'] = (
|
|
self.request_spec.requested_destination)
|
|
|
|
task = self._generate_task()
|
|
gmv_mock.return_value = 23
|
|
|
|
# We just need this hook point to set a uuid on the
|
|
# migration before we use it for teardown
|
|
def set_migration_uuid(*a, **k):
|
|
task._migration.uuid = uuids.migration
|
|
return mock.MagicMock()
|
|
|
|
# NOTE(danms): It's odd to do this on cn_mock, but it's just because
|
|
# of when we need to have it set in the flow and where we have an easy
|
|
# place to find it via self.migration.
|
|
cn_mock.side_effect = set_migration_uuid
|
|
|
|
selection = self.host_lists[0][0]
|
|
with mock.patch.object(task, '_is_selected_host_in_source_cell',
|
|
return_value=same_cell) as _is_source_cell_mock:
|
|
task.execute()
|
|
_is_source_cell_mock.assert_called_once_with(selection)
|
|
|
|
self.ensure_network_information_mock.assert_called_once_with(
|
|
self.instance)
|
|
self.heal_reqspec_is_bfv_mock.assert_called_once_with(
|
|
self.context, self.request_spec, self.instance)
|
|
sig_mock.assert_called_once_with(self.context, self.request_spec)
|
|
task.query_client.select_destinations.assert_called_once_with(
|
|
self.context, self.request_spec, [self.instance.uuid],
|
|
return_objects=True, return_alternates=True)
|
|
|
|
if same_cell:
|
|
prep_resize_mock.assert_called_once_with(
|
|
self.context, self.instance, self.request_spec.image,
|
|
self.flavor, selection.service_host, task._migration,
|
|
request_spec=self.request_spec,
|
|
filter_properties=self.filter_properties,
|
|
node=selection.nodename,
|
|
clean_shutdown=self.clean_shutdown, host_list=[])
|
|
az_mock.assert_called_once_with(self.context, 'host1')
|
|
cross_cell_exec_mock.assert_not_called()
|
|
else:
|
|
cross_cell_exec_mock.assert_called_once_with()
|
|
az_mock.assert_not_called()
|
|
prep_resize_mock.assert_not_called()
|
|
|
|
self.assertIsNotNone(task._migration)
|
|
|
|
old_flavor = self.instance.flavor
|
|
new_flavor = self.flavor
|
|
self.assertEqual(old_flavor.id, task._migration.old_instance_type_id)
|
|
self.assertEqual(new_flavor.id, task._migration.new_instance_type_id)
|
|
self.assertEqual('pre-migrating', task._migration.status)
|
|
self.assertEqual(self.instance.uuid, task._migration.instance_uuid)
|
|
self.assertEqual(self.instance.host, task._migration.source_compute)
|
|
self.assertEqual(self.instance.node, task._migration.source_node)
|
|
if old_flavor.id != new_flavor.id:
|
|
self.assertEqual('resize', task._migration.migration_type)
|
|
else:
|
|
self.assertEqual('migration', task._migration.migration_type)
|
|
|
|
task._migration.create.assert_called_once_with()
|
|
|
|
if requested_destination:
|
|
self.assertIsNone(self.request_spec.retry)
|
|
self.assertIn('cell', self.request_spec.requested_destination)
|
|
self.assertIsNotNone(self.request_spec.requested_destination.cell)
|
|
self.assertEqual(
|
|
not same_cell,
|
|
self.request_spec.requested_destination.allow_cross_cell_move)
|
|
|
|
mock_get_resources.assert_called_once_with(
|
|
self.context, self.instance.uuid)
|
|
self.assertEqual([], self.request_spec.requested_resources)
|
|
|
|
def test_execute(self):
|
|
self._test_execute()
|
|
|
|
def test_execute_with_destination(self):
|
|
self._test_execute(requested_destination=True)
|
|
|
|
def test_execute_resize(self):
|
|
self.flavor = self.flavor.obj_clone()
|
|
self.flavor.id = 3
|
|
self._test_execute()
|
|
|
|
def test_execute_same_cell_false(self):
|
|
"""Tests the execute() scenario that the RequestSpec allows cross
|
|
cell move and the selected target host is in another cell so
|
|
CrossCellMigrationTask is executed.
|
|
"""
|
|
self._test_execute(same_cell=False)
|
|
|
|
@mock.patch.object(objects.MigrationList, 'get_by_filters')
|
|
@mock.patch('nova.conductor.tasks.migrate.revert_allocation_for_migration')
|
|
@mock.patch('nova.scheduler.client.report.SchedulerReportClient')
|
|
@mock.patch('nova.objects.ComputeNode.get_by_host_and_nodename')
|
|
@mock.patch('nova.objects.Migration.save')
|
|
@mock.patch('nova.objects.Migration.create')
|
|
@mock.patch('nova.objects.Service.get_minimum_version_multi')
|
|
@mock.patch('nova.availability_zones.get_host_availability_zone')
|
|
@mock.patch.object(scheduler_utils, 'setup_instance_group')
|
|
@mock.patch.object(query.SchedulerQueryClient, 'select_destinations')
|
|
@mock.patch.object(compute_rpcapi.ComputeAPI, 'prep_resize')
|
|
def test_execute_rollback(self, prep_resize_mock, sel_dest_mock, sig_mock,
|
|
az_mock, gmv_mock, cm_mock, sm_mock, cn_mock,
|
|
rc_mock, mock_ra, mock_gbf):
|
|
sel_dest_mock.return_value = self.host_lists
|
|
az_mock.return_value = 'myaz'
|
|
task = self._generate_task()
|
|
gmv_mock.return_value = 23
|
|
mock_gbf.return_value = objects.MigrationList()
|
|
mock_get_resources = \
|
|
self.mock_network_api.get_requested_resource_for_instance
|
|
mock_get_resources.return_value = []
|
|
|
|
# We just need this hook point to set a uuid on the
|
|
# migration before we use it for teardown
|
|
def set_migration_uuid(*a, **k):
|
|
task._migration.uuid = uuids.migration
|
|
return mock.MagicMock()
|
|
|
|
# NOTE(danms): It's odd to do this on cn_mock, but it's just because
|
|
# of when we need to have it set in the flow and where we have an easy
|
|
# place to find it via self.migration.
|
|
cn_mock.side_effect = set_migration_uuid
|
|
|
|
prep_resize_mock.side_effect = test.TestingException
|
|
task._held_allocations = mock.sentinel.allocs
|
|
self.assertRaises(test.TestingException, task.execute)
|
|
self.assertIsNotNone(task._migration)
|
|
task._migration.create.assert_called_once_with()
|
|
task._migration.save.assert_called_once_with()
|
|
self.assertEqual('error', task._migration.status)
|
|
mock_ra.assert_called_once_with(task.context, task._source_cn,
|
|
task.instance, task._migration)
|
|
mock_get_resources.assert_called_once_with(
|
|
self.context, self.instance.uuid)
|
|
|
|
@mock.patch.object(scheduler_utils, 'claim_resources')
|
|
@mock.patch.object(context.RequestContext, 'elevated')
|
|
def test_execute_reschedule(self, mock_elevated, mock_claim):
|
|
report_client = report.SchedulerReportClient()
|
|
# setup task for re-schedule
|
|
alloc_req = {
|
|
"allocations": {
|
|
uuids.host1: {
|
|
"resources": {
|
|
"VCPU": 1,
|
|
"MEMORY_MB": 1024,
|
|
"DISK_GB": 100}}}}
|
|
alternate_selection = objects.Selection(
|
|
service_host="host1",
|
|
nodename="node1",
|
|
cell_uuid=uuids.cell1,
|
|
allocation_request=jsonutils.dumps(alloc_req),
|
|
allocation_request_version='1.19')
|
|
task = migrate.MigrationTask(
|
|
self.context, self.instance, self.flavor, self.request_spec,
|
|
self.clean_shutdown, compute_rpcapi.ComputeAPI(),
|
|
query.SchedulerQueryClient(), report_client,
|
|
host_list=[alternate_selection], network_api=self.mock_network_api)
|
|
mock_claim.return_value = True
|
|
|
|
actual_selection = task._reschedule()
|
|
|
|
self.assertIs(alternate_selection, actual_selection)
|
|
mock_claim.assert_called_once_with(
|
|
mock_elevated.return_value, report_client, self.request_spec,
|
|
self.instance.uuid, alloc_req, '1.19')
|
|
|
|
@mock.patch.object(scheduler_utils, 'fill_provider_mapping')
|
|
@mock.patch.object(scheduler_utils, 'claim_resources')
|
|
@mock.patch.object(context.RequestContext, 'elevated')
|
|
def test_execute_reschedule_claim_fails_no_more_alternate(
|
|
self, mock_elevated, mock_claim, mock_fill_provider_mapping):
|
|
report_client = report.SchedulerReportClient()
|
|
# set up the task for re-schedule
|
|
alloc_req = {
|
|
"allocations": {
|
|
uuids.host1: {
|
|
"resources": {
|
|
"VCPU": 1,
|
|
"MEMORY_MB": 1024,
|
|
"DISK_GB": 100}}}}
|
|
alternate_selection = objects.Selection(
|
|
service_host="host1",
|
|
nodename="node1",
|
|
cell_uuid=uuids.cell1,
|
|
allocation_request=jsonutils.dumps(alloc_req),
|
|
allocation_request_version='1.19')
|
|
task = migrate.MigrationTask(
|
|
self.context, self.instance, self.flavor, self.request_spec,
|
|
self.clean_shutdown, compute_rpcapi.ComputeAPI(),
|
|
query.SchedulerQueryClient(), report_client,
|
|
host_list=[alternate_selection], network_api=self.mock_network_api)
|
|
mock_claim.return_value = False
|
|
|
|
self.assertRaises(exception.MaxRetriesExceeded, task._reschedule)
|
|
|
|
mock_claim.assert_called_once_with(
|
|
mock_elevated.return_value, report_client, self.request_spec,
|
|
self.instance.uuid, alloc_req, '1.19')
|
|
mock_fill_provider_mapping.assert_not_called()
|
|
|
|
@mock.patch('nova.objects.InstanceMapping.get_by_instance_uuid',
|
|
return_value=objects.InstanceMapping(
|
|
cell_mapping=objects.CellMapping(uuid=uuids.cell1)))
|
|
@mock.patch('nova.conductor.tasks.migrate.LOG.debug')
|
|
def test_set_requested_destination_cell_allow_cross_cell_resize_true(
|
|
self, mock_debug, mock_get_im):
|
|
"""Tests the scenario that the RequestSpec is configured for
|
|
allow_cross_cell_resize=True.
|
|
"""
|
|
task = self._generate_task()
|
|
legacy_props = self.request_spec.to_legacy_filter_properties_dict()
|
|
self.request_spec.requested_destination = objects.Destination(
|
|
allow_cross_cell_move=True)
|
|
task._set_requested_destination_cell(legacy_props)
|
|
mock_get_im.assert_called_once_with(self.context, self.instance.uuid)
|
|
mock_debug.assert_called_once()
|
|
self.assertIn('Allowing migration from cell',
|
|
mock_debug.call_args[0][0])
|
|
self.assertEqual(mock_get_im.return_value.cell_mapping,
|
|
self.request_spec.requested_destination.cell)
|
|
|
|
@mock.patch('nova.objects.InstanceMapping.get_by_instance_uuid',
|
|
return_value=objects.InstanceMapping(
|
|
cell_mapping=objects.CellMapping(uuid=uuids.cell1)))
|
|
@mock.patch('nova.conductor.tasks.migrate.LOG.debug')
|
|
def test_set_requested_destination_cell_allow_cross_cell_resize_true_host(
|
|
self, mock_debug, mock_get_im):
|
|
"""Tests the scenario that the RequestSpec is configured for
|
|
allow_cross_cell_resize=True and there is a requested target host.
|
|
"""
|
|
task = self._generate_task()
|
|
legacy_props = self.request_spec.to_legacy_filter_properties_dict()
|
|
self.request_spec.requested_destination = objects.Destination(
|
|
allow_cross_cell_move=True, host='fake-host')
|
|
task._set_requested_destination_cell(legacy_props)
|
|
mock_get_im.assert_called_once_with(self.context, self.instance.uuid)
|
|
mock_debug.assert_called_once()
|
|
self.assertIn('Not restricting cell', mock_debug.call_args[0][0])
|
|
self.assertIsNone(self.request_spec.requested_destination.cell)
|
|
|
|
@mock.patch('nova.objects.InstanceMapping.get_by_instance_uuid',
|
|
return_value=objects.InstanceMapping(
|
|
cell_mapping=objects.CellMapping(uuid=uuids.cell1)))
|
|
@mock.patch('nova.conductor.tasks.migrate.LOG.debug')
|
|
def test_set_requested_destination_cell_allow_cross_cell_resize_false(
|
|
self, mock_debug, mock_get_im):
|
|
"""Tests the scenario that the RequestSpec is configured for
|
|
allow_cross_cell_resize=False.
|
|
"""
|
|
task = self._generate_task()
|
|
legacy_props = self.request_spec.to_legacy_filter_properties_dict()
|
|
# We don't have to explicitly set RequestSpec.requested_destination
|
|
# since _set_requested_destination_cell will do that and the
|
|
# Destination object will default allow_cross_cell_move to False.
|
|
task._set_requested_destination_cell(legacy_props)
|
|
mock_get_im.assert_called_once_with(self.context, self.instance.uuid)
|
|
mock_debug.assert_called_once()
|
|
self.assertIn('Restricting to cell', mock_debug.call_args[0][0])
|
|
|
|
def test_is_selected_host_in_source_cell_true(self):
|
|
"""Tests the scenario that the host Selection from the scheduler is in
|
|
the same cell as the instance.
|
|
"""
|
|
task = self._generate_task()
|
|
selection = objects.Selection(cell_uuid=self.context.cell_uuid)
|
|
self.assertTrue(task._is_selected_host_in_source_cell(selection))
|
|
|
|
def test_is_selected_host_in_source_cell_false(self):
|
|
"""Tests the scenario that the host Selection from the scheduler is
|
|
not in the same cell as the instance.
|
|
"""
|
|
task = self._generate_task()
|
|
selection = objects.Selection(cell_uuid=uuids.cell2, service_host='x')
|
|
self.assertFalse(task._is_selected_host_in_source_cell(selection))
|
|
|
|
|
|
class MigrationTaskAllocationUtils(test.NoDBTestCase):
|
|
@mock.patch('nova.objects.ComputeNode.get_by_host_and_nodename')
|
|
def test_replace_allocation_with_migration_no_host(self, mock_cn):
|
|
mock_cn.side_effect = exception.ComputeHostNotFound(host='host')
|
|
migration = objects.Migration()
|
|
instance = objects.Instance(host='host', node='node')
|
|
|
|
self.assertRaises(exception.ComputeHostNotFound,
|
|
migrate.replace_allocation_with_migration,
|
|
mock.sentinel.context,
|
|
instance, migration)
|
|
mock_cn.assert_called_once_with(mock.sentinel.context,
|
|
instance.host, instance.node)
|
|
|
|
@mock.patch('nova.scheduler.client.report.SchedulerReportClient.'
|
|
'get_allocs_for_consumer')
|
|
@mock.patch('nova.objects.ComputeNode.get_by_host_and_nodename')
|
|
def test_replace_allocation_with_migration_no_allocs(self, mock_cn,
|
|
mock_ga):
|
|
mock_ga.return_value = {'allocations': {}}
|
|
migration = objects.Migration(uuid=uuids.migration)
|
|
instance = objects.Instance(uuid=uuids.instance,
|
|
host='host', node='node')
|
|
|
|
result = migrate.replace_allocation_with_migration(
|
|
mock.sentinel.context, instance, migration)
|
|
self.assertEqual((None, None), result)
|
|
|
|
@mock.patch('nova.scheduler.client.report.SchedulerReportClient.'
|
|
'put_allocations')
|
|
@mock.patch('nova.scheduler.client.report.SchedulerReportClient.'
|
|
'get_allocs_for_consumer')
|
|
@mock.patch('nova.objects.ComputeNode.get_by_host_and_nodename')
|
|
def test_replace_allocation_with_migration_allocs_fail(self, mock_cn,
|
|
mock_ga, mock_pa):
|
|
ctxt = context.get_admin_context()
|
|
migration = objects.Migration(uuid=uuids.migration)
|
|
instance = objects.Instance(uuid=uuids.instance,
|
|
user_id='fake', project_id='fake',
|
|
host='host', node='node')
|
|
mock_pa.return_value = False
|
|
|
|
self.assertRaises(exception.NoValidHost,
|
|
migrate.replace_allocation_with_migration,
|
|
ctxt,
|
|
instance, migration)
|