670 lines
28 KiB
Python
670 lines
28 KiB
Python
# vim: tabstop=4 shiftwidth=4 softtabstop=4
|
|
|
|
# Copyright 2010 United States Government as represented by the
|
|
# Administrator of the National Aeronautics and Space Administration.
|
|
# All Rights Reserved.
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
|
# not use this file except in compliance with the License. You may obtain
|
|
# a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
|
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
|
# License for the specific language governing permissions and limitations
|
|
# under the License.
|
|
"""
|
|
Tests For Compute
|
|
"""
|
|
|
|
import datetime
|
|
import mox
|
|
|
|
from nova import compute
|
|
from nova import context
|
|
from nova import db
|
|
from nova import exception
|
|
from nova import flags
|
|
from nova import log as logging
|
|
from nova import rpc
|
|
from nova import test
|
|
from nova import utils
|
|
from nova.auth import manager
|
|
from nova.compute import instance_types
|
|
from nova.compute import manager as compute_manager
|
|
from nova.compute import power_state
|
|
from nova.db.sqlalchemy import models
|
|
from nova.image import local
|
|
|
|
LOG = logging.getLogger('nova.tests.compute')
|
|
FLAGS = flags.FLAGS
|
|
flags.DECLARE('stub_network', 'nova.compute.manager')
|
|
flags.DECLARE('live_migration_retry_count', 'nova.compute.manager')
|
|
|
|
|
|
class FakeTime(object):
|
|
def __init__(self):
|
|
self.counter = 0
|
|
|
|
def sleep(self, t):
|
|
self.counter += t
|
|
|
|
|
|
class ComputeTestCase(test.TestCase):
|
|
"""Test case for compute"""
|
|
def setUp(self):
|
|
super(ComputeTestCase, self).setUp()
|
|
self.flags(connection_type='fake',
|
|
stub_network=True,
|
|
network_manager='nova.network.manager.FlatManager')
|
|
self.compute = utils.import_object(FLAGS.compute_manager)
|
|
self.compute_api = compute.API()
|
|
self.manager = manager.AuthManager()
|
|
self.user = self.manager.create_user('fake', 'fake', 'fake')
|
|
self.project = self.manager.create_project('fake', 'fake', 'fake')
|
|
self.context = context.RequestContext('fake', 'fake', False)
|
|
|
|
def fake_show(meh, context, id):
|
|
return {'id': 1, 'properties': {'kernel_id': 1, 'ramdisk_id': 1}}
|
|
|
|
self.stubs.Set(local.LocalImageService, 'show', fake_show)
|
|
|
|
def tearDown(self):
|
|
self.manager.delete_user(self.user)
|
|
self.manager.delete_project(self.project)
|
|
super(ComputeTestCase, self).tearDown()
|
|
|
|
def _create_instance(self, params={}):
|
|
"""Create a test instance"""
|
|
inst = {}
|
|
inst['image_id'] = 1
|
|
inst['reservation_id'] = 'r-fakeres'
|
|
inst['launch_time'] = '10'
|
|
inst['user_id'] = self.user.id
|
|
inst['project_id'] = self.project.id
|
|
type_id = instance_types.get_instance_type_by_name('m1.tiny')['id']
|
|
inst['instance_type_id'] = type_id
|
|
inst['mac_address'] = utils.generate_mac()
|
|
inst['ami_launch_index'] = 0
|
|
inst.update(params)
|
|
return db.instance_create(self.context, inst)['id']
|
|
|
|
def _create_instance_type(self, params={}):
|
|
"""Create a test instance"""
|
|
context = self.context.elevated()
|
|
inst = {}
|
|
inst['name'] = 'm1.small'
|
|
inst['memory_mb'] = '1024'
|
|
inst['vcpus'] = '1'
|
|
inst['local_gb'] = '20'
|
|
inst['flavorid'] = '1'
|
|
inst['swap'] = '2048'
|
|
inst['rxtx_quota'] = 100
|
|
inst['rxtx_cap'] = 200
|
|
inst.update(params)
|
|
return db.instance_type_create(context, inst)['id']
|
|
|
|
def _create_group(self):
|
|
values = {'name': 'testgroup',
|
|
'description': 'testgroup',
|
|
'user_id': self.user.id,
|
|
'project_id': self.project.id}
|
|
return db.security_group_create(self.context, values)
|
|
|
|
def _get_dummy_instance(self):
|
|
"""Get mock-return-value instance object
|
|
Use this when any testcase executed later than test_run_terminate
|
|
"""
|
|
vol1 = models.Volume()
|
|
vol1['id'] = 1
|
|
vol2 = models.Volume()
|
|
vol2['id'] = 2
|
|
instance_ref = models.Instance()
|
|
instance_ref['id'] = 1
|
|
instance_ref['volumes'] = [vol1, vol2]
|
|
instance_ref['hostname'] = 'i-00000001'
|
|
instance_ref['host'] = 'dummy'
|
|
return instance_ref
|
|
|
|
def test_create_instance_defaults_display_name(self):
|
|
"""Verify that an instance cannot be created without a display_name."""
|
|
cases = [dict(), dict(display_name=None)]
|
|
for instance in cases:
|
|
ref = self.compute_api.create(self.context,
|
|
instance_types.get_default_instance_type(), None, **instance)
|
|
try:
|
|
self.assertNotEqual(ref[0]['display_name'], None)
|
|
finally:
|
|
db.instance_destroy(self.context, ref[0]['id'])
|
|
|
|
def test_create_instance_associates_security_groups(self):
|
|
"""Make sure create associates security groups"""
|
|
group = self._create_group()
|
|
ref = self.compute_api.create(
|
|
self.context,
|
|
instance_type=instance_types.get_default_instance_type(),
|
|
image_id=None,
|
|
security_group=['testgroup'])
|
|
try:
|
|
self.assertEqual(len(db.security_group_get_by_instance(
|
|
self.context, ref[0]['id'])), 1)
|
|
group = db.security_group_get(self.context, group['id'])
|
|
self.assert_(len(group.instances) == 1)
|
|
finally:
|
|
db.security_group_destroy(self.context, group['id'])
|
|
db.instance_destroy(self.context, ref[0]['id'])
|
|
|
|
def test_destroy_instance_disassociates_security_groups(self):
|
|
"""Make sure destroying disassociates security groups"""
|
|
group = self._create_group()
|
|
|
|
ref = self.compute_api.create(
|
|
self.context,
|
|
instance_type=instance_types.get_default_instance_type(),
|
|
image_id=None,
|
|
security_group=['testgroup'])
|
|
try:
|
|
db.instance_destroy(self.context, ref[0]['id'])
|
|
group = db.security_group_get(self.context, group['id'])
|
|
self.assert_(len(group.instances) == 0)
|
|
finally:
|
|
db.security_group_destroy(self.context, group['id'])
|
|
|
|
def test_destroy_security_group_disassociates_instances(self):
|
|
"""Make sure destroying security groups disassociates instances"""
|
|
group = self._create_group()
|
|
|
|
ref = self.compute_api.create(
|
|
self.context,
|
|
instance_type=instance_types.get_default_instance_type(),
|
|
image_id=None,
|
|
security_group=['testgroup'])
|
|
|
|
try:
|
|
db.security_group_destroy(self.context, group['id'])
|
|
group = db.security_group_get(context.get_admin_context(
|
|
read_deleted=True), group['id'])
|
|
self.assert_(len(group.instances) == 0)
|
|
finally:
|
|
db.instance_destroy(self.context, ref[0]['id'])
|
|
|
|
def test_run_terminate(self):
|
|
"""Make sure it is possible to run and terminate instance"""
|
|
instance_id = self._create_instance()
|
|
|
|
self.compute.run_instance(self.context, instance_id)
|
|
|
|
instances = db.instance_get_all(context.get_admin_context())
|
|
LOG.info(_("Running instances: %s"), instances)
|
|
self.assertEqual(len(instances), 1)
|
|
|
|
self.compute.terminate_instance(self.context, instance_id)
|
|
|
|
instances = db.instance_get_all(context.get_admin_context())
|
|
LOG.info(_("After terminating instances: %s"), instances)
|
|
self.assertEqual(len(instances), 0)
|
|
|
|
def test_run_terminate_timestamps(self):
|
|
"""Make sure timestamps are set for launched and destroyed"""
|
|
instance_id = self._create_instance()
|
|
instance_ref = db.instance_get(self.context, instance_id)
|
|
self.assertEqual(instance_ref['launched_at'], None)
|
|
self.assertEqual(instance_ref['deleted_at'], None)
|
|
launch = datetime.datetime.utcnow()
|
|
self.compute.run_instance(self.context, instance_id)
|
|
instance_ref = db.instance_get(self.context, instance_id)
|
|
self.assert_(instance_ref['launched_at'] > launch)
|
|
self.assertEqual(instance_ref['deleted_at'], None)
|
|
terminate = datetime.datetime.utcnow()
|
|
self.compute.terminate_instance(self.context, instance_id)
|
|
self.context = self.context.elevated(True)
|
|
instance_ref = db.instance_get(self.context, instance_id)
|
|
self.assert_(instance_ref['launched_at'] < terminate)
|
|
self.assert_(instance_ref['deleted_at'] > terminate)
|
|
|
|
def test_pause(self):
|
|
"""Ensure instance can be paused"""
|
|
instance_id = self._create_instance()
|
|
self.compute.run_instance(self.context, instance_id)
|
|
self.compute.pause_instance(self.context, instance_id)
|
|
self.compute.unpause_instance(self.context, instance_id)
|
|
self.compute.terminate_instance(self.context, instance_id)
|
|
|
|
def test_suspend(self):
|
|
"""ensure instance can be suspended"""
|
|
instance_id = self._create_instance()
|
|
self.compute.run_instance(self.context, instance_id)
|
|
self.compute.suspend_instance(self.context, instance_id)
|
|
self.compute.resume_instance(self.context, instance_id)
|
|
self.compute.terminate_instance(self.context, instance_id)
|
|
|
|
def test_reboot(self):
|
|
"""Ensure instance can be rebooted"""
|
|
instance_id = self._create_instance()
|
|
self.compute.run_instance(self.context, instance_id)
|
|
self.compute.reboot_instance(self.context, instance_id)
|
|
self.compute.terminate_instance(self.context, instance_id)
|
|
|
|
def test_set_admin_password(self):
|
|
"""Ensure instance can have its admin password set"""
|
|
instance_id = self._create_instance()
|
|
self.compute.run_instance(self.context, instance_id)
|
|
self.compute.set_admin_password(self.context, instance_id)
|
|
self.compute.terminate_instance(self.context, instance_id)
|
|
|
|
def test_inject_file(self):
|
|
"""Ensure we can write a file to an instance"""
|
|
instance_id = self._create_instance()
|
|
self.compute.run_instance(self.context, instance_id)
|
|
self.compute.inject_file(self.context, instance_id, "/tmp/test",
|
|
"File Contents")
|
|
self.compute.terminate_instance(self.context, instance_id)
|
|
|
|
def test_snapshot(self):
|
|
"""Ensure instance can be snapshotted"""
|
|
instance_id = self._create_instance()
|
|
name = "myfakesnapshot"
|
|
self.compute.run_instance(self.context, instance_id)
|
|
self.compute.snapshot_instance(self.context, instance_id, name)
|
|
self.compute.terminate_instance(self.context, instance_id)
|
|
|
|
def test_console_output(self):
|
|
"""Make sure we can get console output from instance"""
|
|
instance_id = self._create_instance()
|
|
self.compute.run_instance(self.context, instance_id)
|
|
|
|
console = self.compute.get_console_output(self.context,
|
|
instance_id)
|
|
self.assert_(console)
|
|
self.compute.terminate_instance(self.context, instance_id)
|
|
|
|
def test_ajax_console(self):
|
|
"""Make sure we can get console output from instance"""
|
|
instance_id = self._create_instance()
|
|
self.compute.run_instance(self.context, instance_id)
|
|
|
|
console = self.compute.get_ajax_console(self.context,
|
|
instance_id)
|
|
self.assert_(set(['token', 'host', 'port']).issubset(console.keys()))
|
|
self.compute.terminate_instance(self.context, instance_id)
|
|
|
|
def test_vnc_console(self):
|
|
"""Make sure we can a vnc console for an instance."""
|
|
instance_id = self._create_instance()
|
|
self.compute.run_instance(self.context, instance_id)
|
|
|
|
console = self.compute.get_vnc_console(self.context,
|
|
instance_id)
|
|
self.assert_(console)
|
|
self.compute.terminate_instance(self.context, instance_id)
|
|
|
|
def test_run_instance_existing(self):
|
|
"""Ensure failure when running an instance that already exists"""
|
|
instance_id = self._create_instance()
|
|
self.compute.run_instance(self.context, instance_id)
|
|
self.assertRaises(exception.Error,
|
|
self.compute.run_instance,
|
|
self.context,
|
|
instance_id)
|
|
self.compute.terminate_instance(self.context, instance_id)
|
|
|
|
def test_lock(self):
|
|
"""ensure locked instance cannot be changed"""
|
|
instance_id = self._create_instance()
|
|
self.compute.run_instance(self.context, instance_id)
|
|
|
|
non_admin_context = context.RequestContext(None, None, False, False)
|
|
|
|
# decorator should return False (fail) with locked nonadmin context
|
|
self.compute.lock_instance(self.context, instance_id)
|
|
ret_val = self.compute.reboot_instance(non_admin_context, instance_id)
|
|
self.assertEqual(ret_val, False)
|
|
|
|
# decorator should return None (success) with unlocked nonadmin context
|
|
self.compute.unlock_instance(self.context, instance_id)
|
|
ret_val = self.compute.reboot_instance(non_admin_context, instance_id)
|
|
self.assertEqual(ret_val, None)
|
|
|
|
self.compute.terminate_instance(self.context, instance_id)
|
|
|
|
def test_resize_instance(self):
|
|
"""Ensure instance can be migrated/resized"""
|
|
instance_id = self._create_instance()
|
|
context = self.context.elevated()
|
|
|
|
self.compute.run_instance(self.context, instance_id)
|
|
db.instance_update(self.context, instance_id, {'host': 'foo'})
|
|
self.compute.prep_resize(context, instance_id, 1)
|
|
migration_ref = db.migration_get_by_instance_and_status(context,
|
|
instance_id, 'pre-migrating')
|
|
self.compute.resize_instance(context, instance_id,
|
|
migration_ref['id'])
|
|
self.compute.terminate_instance(context, instance_id)
|
|
|
|
def test_resize_invalid_flavor_fails(self):
|
|
"""Ensure invalid flavors raise"""
|
|
instance_id = self._create_instance()
|
|
context = self.context.elevated()
|
|
self.compute.run_instance(self.context, instance_id)
|
|
|
|
self.assertRaises(exception.NotFound, self.compute_api.resize,
|
|
context, instance_id, 200)
|
|
|
|
self.compute.terminate_instance(context, instance_id)
|
|
|
|
def test_resize_down_fails(self):
|
|
"""Ensure resizing down raises and fails"""
|
|
context = self.context.elevated()
|
|
instance_id = self._create_instance()
|
|
|
|
self.compute.run_instance(self.context, instance_id)
|
|
inst_type = instance_types.get_instance_type_by_name('m1.xlarge')
|
|
db.instance_update(self.context, instance_id,
|
|
{'instance_type_id': inst_type['id']})
|
|
|
|
self.assertRaises(exception.ApiError, self.compute_api.resize,
|
|
context, instance_id, 1)
|
|
|
|
self.compute.terminate_instance(context, instance_id)
|
|
|
|
def test_resize_same_size_fails(self):
|
|
"""Ensure invalid flavors raise"""
|
|
context = self.context.elevated()
|
|
instance_id = self._create_instance()
|
|
|
|
self.compute.run_instance(self.context, instance_id)
|
|
|
|
self.assertRaises(exception.ApiError, self.compute_api.resize,
|
|
context, instance_id, 1)
|
|
|
|
self.compute.terminate_instance(context, instance_id)
|
|
|
|
def test_get_by_flavor_id(self):
|
|
type = instance_types.get_instance_type_by_flavor_id(1)
|
|
self.assertEqual(type['name'], 'm1.tiny')
|
|
|
|
def test_resize_same_source_fails(self):
|
|
"""Ensure instance fails to migrate when source and destination are
|
|
the same host"""
|
|
instance_id = self._create_instance()
|
|
self.compute.run_instance(self.context, instance_id)
|
|
self.assertRaises(exception.Error, self.compute.prep_resize,
|
|
self.context, instance_id, 1)
|
|
self.compute.terminate_instance(self.context, instance_id)
|
|
|
|
def _setup_other_managers(self):
|
|
self.volume_manager = utils.import_object(FLAGS.volume_manager)
|
|
self.network_manager = utils.import_object(FLAGS.network_manager)
|
|
self.compute_driver = utils.import_object(FLAGS.compute_driver)
|
|
|
|
def test_pre_live_migration_instance_has_no_fixed_ip(self):
|
|
"""Confirm raising exception if instance doesn't have fixed_ip."""
|
|
instance_ref = self._get_dummy_instance()
|
|
c = context.get_admin_context()
|
|
i_id = instance_ref['id']
|
|
|
|
dbmock = self.mox.CreateMock(db)
|
|
dbmock.instance_get(c, i_id).AndReturn(instance_ref)
|
|
dbmock.instance_get_fixed_address(c, i_id).AndReturn(None)
|
|
|
|
self.compute.db = dbmock
|
|
self.mox.ReplayAll()
|
|
self.assertRaises(exception.NotFound,
|
|
self.compute.pre_live_migration,
|
|
c, instance_ref['id'], time=FakeTime())
|
|
|
|
def test_pre_live_migration_instance_has_volume(self):
|
|
"""Confirm setup_compute_volume is called when volume is mounted."""
|
|
i_ref = self._get_dummy_instance()
|
|
c = context.get_admin_context()
|
|
|
|
self._setup_other_managers()
|
|
dbmock = self.mox.CreateMock(db)
|
|
volmock = self.mox.CreateMock(self.volume_manager)
|
|
netmock = self.mox.CreateMock(self.network_manager)
|
|
drivermock = self.mox.CreateMock(self.compute_driver)
|
|
|
|
dbmock.instance_get(c, i_ref['id']).AndReturn(i_ref)
|
|
dbmock.instance_get_fixed_address(c, i_ref['id']).AndReturn('dummy')
|
|
for i in range(len(i_ref['volumes'])):
|
|
vid = i_ref['volumes'][i]['id']
|
|
volmock.setup_compute_volume(c, vid).InAnyOrder('g1')
|
|
netmock.setup_compute_network(c, i_ref['id'])
|
|
drivermock.ensure_filtering_rules_for_instance(i_ref)
|
|
|
|
self.compute.db = dbmock
|
|
self.compute.volume_manager = volmock
|
|
self.compute.network_manager = netmock
|
|
self.compute.driver = drivermock
|
|
|
|
self.mox.ReplayAll()
|
|
ret = self.compute.pre_live_migration(c, i_ref['id'])
|
|
self.assertEqual(ret, None)
|
|
|
|
def test_pre_live_migration_instance_has_no_volume(self):
|
|
"""Confirm log meg when instance doesn't mount any volumes."""
|
|
i_ref = self._get_dummy_instance()
|
|
i_ref['volumes'] = []
|
|
c = context.get_admin_context()
|
|
|
|
self._setup_other_managers()
|
|
dbmock = self.mox.CreateMock(db)
|
|
netmock = self.mox.CreateMock(self.network_manager)
|
|
drivermock = self.mox.CreateMock(self.compute_driver)
|
|
|
|
dbmock.instance_get(c, i_ref['id']).AndReturn(i_ref)
|
|
dbmock.instance_get_fixed_address(c, i_ref['id']).AndReturn('dummy')
|
|
self.mox.StubOutWithMock(compute_manager.LOG, 'info')
|
|
compute_manager.LOG.info(_("%s has no volume."), i_ref['hostname'])
|
|
netmock.setup_compute_network(c, i_ref['id'])
|
|
drivermock.ensure_filtering_rules_for_instance(i_ref)
|
|
|
|
self.compute.db = dbmock
|
|
self.compute.network_manager = netmock
|
|
self.compute.driver = drivermock
|
|
|
|
self.mox.ReplayAll()
|
|
ret = self.compute.pre_live_migration(c, i_ref['id'], time=FakeTime())
|
|
self.assertEqual(ret, None)
|
|
|
|
def test_pre_live_migration_setup_compute_node_fail(self):
|
|
"""Confirm operation setup_compute_network() fails.
|
|
|
|
It retries and raise exception when timeout exceeded.
|
|
|
|
"""
|
|
|
|
i_ref = self._get_dummy_instance()
|
|
c = context.get_admin_context()
|
|
|
|
self._setup_other_managers()
|
|
dbmock = self.mox.CreateMock(db)
|
|
netmock = self.mox.CreateMock(self.network_manager)
|
|
volmock = self.mox.CreateMock(self.volume_manager)
|
|
|
|
dbmock.instance_get(c, i_ref['id']).AndReturn(i_ref)
|
|
dbmock.instance_get_fixed_address(c, i_ref['id']).AndReturn('dummy')
|
|
for i in range(len(i_ref['volumes'])):
|
|
volmock.setup_compute_volume(c, i_ref['volumes'][i]['id'])
|
|
for i in range(FLAGS.live_migration_retry_count):
|
|
netmock.setup_compute_network(c, i_ref['id']).\
|
|
AndRaise(exception.ProcessExecutionError())
|
|
|
|
self.compute.db = dbmock
|
|
self.compute.network_manager = netmock
|
|
self.compute.volume_manager = volmock
|
|
|
|
self.mox.ReplayAll()
|
|
self.assertRaises(exception.ProcessExecutionError,
|
|
self.compute.pre_live_migration,
|
|
c, i_ref['id'], time=FakeTime())
|
|
|
|
def test_live_migration_works_correctly_with_volume(self):
|
|
"""Confirm check_for_export to confirm volume health check."""
|
|
i_ref = self._get_dummy_instance()
|
|
c = context.get_admin_context()
|
|
topic = db.queue_get_for(c, FLAGS.compute_topic, i_ref['host'])
|
|
|
|
dbmock = self.mox.CreateMock(db)
|
|
dbmock.instance_get(c, i_ref['id']).AndReturn(i_ref)
|
|
self.mox.StubOutWithMock(rpc, 'call')
|
|
rpc.call(c, FLAGS.volume_topic, {"method": "check_for_export",
|
|
"args": {'instance_id': i_ref['id']}})
|
|
dbmock.queue_get_for(c, FLAGS.compute_topic, i_ref['host']).\
|
|
AndReturn(topic)
|
|
rpc.call(c, topic, {"method": "pre_live_migration",
|
|
"args": {'instance_id': i_ref['id']}})
|
|
self.mox.StubOutWithMock(self.compute.driver, 'live_migration')
|
|
self.compute.driver.live_migration(c, i_ref, i_ref['host'],
|
|
self.compute.post_live_migration,
|
|
self.compute.recover_live_migration)
|
|
|
|
self.compute.db = dbmock
|
|
self.mox.ReplayAll()
|
|
ret = self.compute.live_migration(c, i_ref['id'], i_ref['host'])
|
|
self.assertEqual(ret, None)
|
|
|
|
def test_live_migration_dest_raises_exception(self):
|
|
"""Confirm exception when pre_live_migration fails."""
|
|
i_ref = self._get_dummy_instance()
|
|
c = context.get_admin_context()
|
|
topic = db.queue_get_for(c, FLAGS.compute_topic, i_ref['host'])
|
|
|
|
dbmock = self.mox.CreateMock(db)
|
|
dbmock.instance_get(c, i_ref['id']).AndReturn(i_ref)
|
|
self.mox.StubOutWithMock(rpc, 'call')
|
|
rpc.call(c, FLAGS.volume_topic, {"method": "check_for_export",
|
|
"args": {'instance_id': i_ref['id']}})
|
|
dbmock.queue_get_for(c, FLAGS.compute_topic, i_ref['host']).\
|
|
AndReturn(topic)
|
|
rpc.call(c, topic, {"method": "pre_live_migration",
|
|
"args": {'instance_id': i_ref['id']}}).\
|
|
AndRaise(rpc.RemoteError('', '', ''))
|
|
dbmock.instance_update(c, i_ref['id'], {'state_description': 'running',
|
|
'state': power_state.RUNNING,
|
|
'host': i_ref['host']})
|
|
for v in i_ref['volumes']:
|
|
dbmock.volume_update(c, v['id'], {'status': 'in-use'})
|
|
|
|
self.compute.db = dbmock
|
|
self.mox.ReplayAll()
|
|
self.assertRaises(rpc.RemoteError,
|
|
self.compute.live_migration,
|
|
c, i_ref['id'], i_ref['host'])
|
|
|
|
def test_live_migration_dest_raises_exception_no_volume(self):
|
|
"""Same as above test(input pattern is different) """
|
|
i_ref = self._get_dummy_instance()
|
|
i_ref['volumes'] = []
|
|
c = context.get_admin_context()
|
|
topic = db.queue_get_for(c, FLAGS.compute_topic, i_ref['host'])
|
|
|
|
dbmock = self.mox.CreateMock(db)
|
|
dbmock.instance_get(c, i_ref['id']).AndReturn(i_ref)
|
|
dbmock.queue_get_for(c, FLAGS.compute_topic, i_ref['host']).\
|
|
AndReturn(topic)
|
|
self.mox.StubOutWithMock(rpc, 'call')
|
|
rpc.call(c, topic, {"method": "pre_live_migration",
|
|
"args": {'instance_id': i_ref['id']}}).\
|
|
AndRaise(rpc.RemoteError('', '', ''))
|
|
dbmock.instance_update(c, i_ref['id'], {'state_description': 'running',
|
|
'state': power_state.RUNNING,
|
|
'host': i_ref['host']})
|
|
|
|
self.compute.db = dbmock
|
|
self.mox.ReplayAll()
|
|
self.assertRaises(rpc.RemoteError,
|
|
self.compute.live_migration,
|
|
c, i_ref['id'], i_ref['host'])
|
|
|
|
def test_live_migration_works_correctly_no_volume(self):
|
|
"""Confirm live_migration() works as expected correctly."""
|
|
i_ref = self._get_dummy_instance()
|
|
i_ref['volumes'] = []
|
|
c = context.get_admin_context()
|
|
topic = db.queue_get_for(c, FLAGS.compute_topic, i_ref['host'])
|
|
|
|
dbmock = self.mox.CreateMock(db)
|
|
dbmock.instance_get(c, i_ref['id']).AndReturn(i_ref)
|
|
self.mox.StubOutWithMock(rpc, 'call')
|
|
dbmock.queue_get_for(c, FLAGS.compute_topic, i_ref['host']).\
|
|
AndReturn(topic)
|
|
rpc.call(c, topic, {"method": "pre_live_migration",
|
|
"args": {'instance_id': i_ref['id']}})
|
|
self.mox.StubOutWithMock(self.compute.driver, 'live_migration')
|
|
self.compute.driver.live_migration(c, i_ref, i_ref['host'],
|
|
self.compute.post_live_migration,
|
|
self.compute.recover_live_migration)
|
|
|
|
self.compute.db = dbmock
|
|
self.mox.ReplayAll()
|
|
ret = self.compute.live_migration(c, i_ref['id'], i_ref['host'])
|
|
self.assertEqual(ret, None)
|
|
|
|
def test_post_live_migration_working_correctly(self):
|
|
"""Confirm post_live_migration() works as expected correctly."""
|
|
dest = 'desthost'
|
|
flo_addr = '1.2.1.2'
|
|
|
|
# Preparing datas
|
|
c = context.get_admin_context()
|
|
instance_id = self._create_instance()
|
|
i_ref = db.instance_get(c, instance_id)
|
|
db.instance_update(c, i_ref['id'], {'state_description': 'migrating',
|
|
'state': power_state.PAUSED})
|
|
v_ref = db.volume_create(c, {'size': 1, 'instance_id': instance_id})
|
|
fix_addr = db.fixed_ip_create(c, {'address': '1.1.1.1',
|
|
'instance_id': instance_id})
|
|
fix_ref = db.fixed_ip_get_by_address(c, fix_addr)
|
|
flo_ref = db.floating_ip_create(c, {'address': flo_addr,
|
|
'fixed_ip_id': fix_ref['id']})
|
|
# reload is necessary before setting mocks
|
|
i_ref = db.instance_get(c, instance_id)
|
|
|
|
# Preparing mocks
|
|
self.mox.StubOutWithMock(self.compute.volume_manager,
|
|
'remove_compute_volume')
|
|
for v in i_ref['volumes']:
|
|
self.compute.volume_manager.remove_compute_volume(c, v['id'])
|
|
self.mox.StubOutWithMock(self.compute.driver, 'unfilter_instance')
|
|
self.compute.driver.unfilter_instance(i_ref)
|
|
|
|
# executing
|
|
self.mox.ReplayAll()
|
|
ret = self.compute.post_live_migration(c, i_ref, dest)
|
|
|
|
# make sure every data is rewritten to dest
|
|
i_ref = db.instance_get(c, i_ref['id'])
|
|
c1 = (i_ref['host'] == dest)
|
|
flo_refs = db.floating_ip_get_all_by_host(c, dest)
|
|
c2 = (len(flo_refs) != 0 and flo_refs[0]['address'] == flo_addr)
|
|
|
|
# post operaton
|
|
self.assertTrue(c1 and c2)
|
|
db.instance_destroy(c, instance_id)
|
|
db.volume_destroy(c, v_ref['id'])
|
|
db.floating_ip_destroy(c, flo_addr)
|
|
|
|
def test_run_kill_vm(self):
|
|
"""Detect when a vm is terminated behind the scenes"""
|
|
instance_id = self._create_instance()
|
|
|
|
self.compute.run_instance(self.context, instance_id)
|
|
|
|
instances = db.instance_get_all(context.get_admin_context())
|
|
LOG.info(_("Running instances: %s"), instances)
|
|
self.assertEqual(len(instances), 1)
|
|
|
|
instance_name = instances[0].name
|
|
self.compute.driver.test_remove_vm(instance_name)
|
|
|
|
# Force the compute manager to do its periodic poll
|
|
error_list = self.compute.periodic_tasks(context.get_admin_context())
|
|
self.assertFalse(error_list)
|
|
|
|
instances = db.instance_get_all(context.get_admin_context())
|
|
LOG.info(_("After force-killing instances: %s"), instances)
|
|
self.assertEqual(len(instances), 0)
|