diff options
author | Chris Behrens <cbehrens@codestud.com> | 2014-05-20 14:55:13 -0700 |
---|---|---|
committer | Michael Still <mikal@stillhq.com> | 2014-07-22 19:37:12 +1000 |
commit | 3883697d31809ab6c5c448bf853bfde69ca53db0 (patch) | |
tree | 6df2a0af4822092af9af691302b39c4d91b1c444 | |
parent | a905cf2c0cacfc5d0503b3a4ac19ef21eb32e287 (diff) | |
download | nova-3883697d31809ab6c5c448bf853bfde69ca53db0.tar.gz |
Fix last of direct use of object modules
This replaces all uses of nova.objects.<module>.<object> with
nova.objects.<object> in the remaining places.
Implements-Blueprint: object-subclassing
Change-Id: Ic7632cca2455a38abcbdb94feb7e39cfb898bb27
-rw-r--r-- | nova/cells/messaging.py | 3 | ||||
-rw-r--r-- | nova/cells/scheduler.py | 3 | ||||
-rw-r--r-- | nova/cmd/dhcpbridge.py | 3 | ||||
-rw-r--r-- | nova/conductor/manager.py | 7 | ||||
-rw-r--r-- | nova/quota.py | 9 | ||||
-rw-r--r-- | nova/scheduler/filter_scheduler.py | 5 | ||||
-rw-r--r-- | nova/tests/cells/test_cells_messaging.py | 6 | ||||
-rw-r--r-- | nova/tests/fake_instance.py | 7 | ||||
-rw-r--r-- | nova/tests/scheduler/test_filter_scheduler.py | 11 | ||||
-rw-r--r-- | nova/tests/virt/baremetal/test_pxe.py | 24 | ||||
-rw-r--r-- | nova/tests/virt/libvirt/test_driver.py | 71 | ||||
-rw-r--r-- | nova/tests/virt/test_virt_drivers.py | 13 | ||||
-rw-r--r-- | nova/virt/baremetal/pxe.py | 10 | ||||
-rw-r--r-- | nova/virt/firewall.py | 8 | ||||
-rw-r--r-- | nova/virt/libvirt/driver.py | 10 | ||||
-rw-r--r-- | nova/virt/xenapi/host.py | 5 |
16 files changed, 89 insertions, 106 deletions
diff --git a/nova/cells/messaging.py b/nova/cells/messaging.py index 55fbfd75f1..64236f4712 100644 --- a/nova/cells/messaging.py +++ b/nova/cells/messaging.py @@ -48,7 +48,6 @@ from nova.i18n import _ from nova.network import model as network_model from nova import objects from nova.objects import base as objects_base -from nova.objects import instance_fault as instance_fault_obj from nova.openstack.common import excutils from nova.openstack.common import importutils from nova.openstack.common import jsonutils @@ -1103,7 +1102,7 @@ class _BroadcastMessageMethods(_BaseMessageMethods): log_str = _("Got message to create instance fault: " "%(instance_fault)s") LOG.debug(log_str, {'instance_fault': instance_fault}) - fault = instance_fault_obj.InstanceFault(context=message.ctxt) + fault = objects.InstanceFault(context=message.ctxt) fault.update(instance_fault) fault.create() diff --git a/nova/cells/scheduler.py b/nova/cells/scheduler.py index d9552f4324..42d4ff9092 100644 --- a/nova/cells/scheduler.py +++ b/nova/cells/scheduler.py @@ -33,7 +33,6 @@ from nova import exception from nova.i18n import _ from nova import objects from nova.objects import base as obj_base -from nova.objects import instance_action as instance_action_obj from nova.openstack.common import log as logging from nova.scheduler import utils as scheduler_utils from nova import utils @@ -120,7 +119,7 @@ class CellsScheduler(base.Base): def _create_action_here(self, ctxt, instance_uuids): for instance_uuid in instance_uuids: - instance_action_obj.InstanceAction.action_start( + objects.InstanceAction.action_start( ctxt, instance_uuid, instance_actions.CREATE, diff --git a/nova/cmd/dhcpbridge.py b/nova/cmd/dhcpbridge.py index 2abb6a8ffd..114b7484a2 100644 --- a/nova/cmd/dhcpbridge.py +++ b/nova/cmd/dhcpbridge.py @@ -35,7 +35,6 @@ from nova.i18n import _ from nova.network import rpcapi as network_rpcapi from nova import objects from nova.objects import base as objects_base -from nova.objects import network as network_obj from nova.openstack.common import importutils from nova.openstack.common import jsonutils from nova.openstack.common import log as logging @@ -72,7 +71,7 @@ def del_lease(mac, ip_address): def init_leases(network_id): """Get the list of hosts for a network.""" ctxt = context.get_admin_context() - network = network_obj.Network.get_by_id(ctxt, network_id) + network = objects.Network.get_by_id(ctxt, network_id) network_manager = importutils.import_object(CONF.network_manager) return network_manager.get_dhcp_leases(ctxt, network) diff --git a/nova/conductor/manager.py b/nova/conductor/manager.py index 97fb87c849..ba102f4e3d 100644 --- a/nova/conductor/manager.py +++ b/nova/conductor/manager.py @@ -39,7 +39,6 @@ from nova.network.security_group import openstack_driver from nova import notifications from nova import objects from nova.objects import base as nova_object -from nova.objects import quotas as quotas_obj from nova.openstack.common import excutils from nova.openstack.common import jsonutils from nova.openstack.common import log as logging @@ -503,9 +502,9 @@ class ComputeTaskManager(base.Base): request_spec = scheduler_utils.build_request_spec( context, image, [instance], instance_type=flavor) - quotas = quotas_obj.Quotas.from_reservations(context, - reservations, - instance=instance) + quotas = objects.Quotas.from_reservations(context, + reservations, + instance=instance) try: scheduler_utils.populate_retry(filter_properties, instance['uuid']) hosts = self.scheduler_rpcapi.select_destinations( diff --git a/nova/quota.py b/nova/quota.py index 4e1644a178..98cf7fe9dd 100644 --- a/nova/quota.py +++ b/nova/quota.py @@ -24,7 +24,7 @@ import six from nova import db from nova import exception from nova.i18n import _ -from nova.objects import keypair as keypair_obj +from nova import objects from nova.openstack.common import importutils from nova.openstack.common import log as logging from nova.openstack.common import timeutils @@ -1405,6 +1405,11 @@ class QuotaEngine(object): return sorted(self._resources.keys()) +def _keypair_get_count_by_user(*args, **kwargs): + """Helper method to avoid referencing objects.KeyPairList on import.""" + return objects.KeyPairList.get_count_by_user(*args, **kwargs) + + QUOTAS = QuotaEngine() @@ -1426,7 +1431,7 @@ resources = [ CountableResource('security_group_rules', db.security_group_rule_count_by_group, 'quota_security_group_rules'), - CountableResource('key_pairs', keypair_obj.KeyPairList.get_count_by_user, + CountableResource('key_pairs', _keypair_get_count_by_user, 'quota_key_pairs'), ] diff --git a/nova/scheduler/filter_scheduler.py b/nova/scheduler/filter_scheduler.py index 415f46d32d..35220ea64c 100644 --- a/nova/scheduler/filter_scheduler.py +++ b/nova/scheduler/filter_scheduler.py @@ -26,7 +26,7 @@ from oslo.config import cfg from nova.compute import rpcapi as compute_rpcapi from nova import exception from nova.i18n import _ -from nova.objects import instance_group as instance_group_obj +from nova import objects from nova.openstack.common import log as logging from nova.pci import pci_request from nova import rpc @@ -208,8 +208,7 @@ class FilterScheduler(driver.Scheduler): scheduler_hints = filter_properties.get('scheduler_hints') or {} group_hint = scheduler_hints.get('group', None) if group_hint: - group = instance_group_obj.InstanceGroup.get_by_hint(context, - group_hint) + group = objects.InstanceGroup.get_by_hint(context, group_hint) policies = set(('anti-affinity', 'affinity')) if any((policy in policies) for policy in group.policies): update_group_hosts = True diff --git a/nova/tests/cells/test_cells_messaging.py b/nova/tests/cells/test_cells_messaging.py index 3acd49bdc4..e4fdfde7f3 100644 --- a/nova/tests/cells/test_cells_messaging.py +++ b/nova/tests/cells/test_cells_messaging.py @@ -33,7 +33,6 @@ from nova.network import model as network_model from nova import objects from nova.objects import base as objects_base from nova.objects import fields as objects_fields -from nova.objects import instance_fault as instance_fault_obj from nova.openstack.common import jsonutils from nova.openstack.common import timeutils from nova.openstack.common import uuidutils @@ -1621,7 +1620,7 @@ class CellsBroadcastMethodsTestCase(test.TestCase): 'message': 'fake-message', 'details': 'fake-details'} - if_mock = mock.Mock(spec_set=instance_fault_obj.InstanceFault) + if_mock = mock.Mock(spec_set=objects.InstanceFault) def _check_create(): self.assertEqual('fake-message', if_mock.message) @@ -1631,8 +1630,7 @@ class CellsBroadcastMethodsTestCase(test.TestCase): if_mock.create.side_effect = _check_create - with mock.patch.object(instance_fault_obj, - 'InstanceFault') as if_obj_mock: + with mock.patch.object(objects, 'InstanceFault') as if_obj_mock: if_obj_mock.return_value = if_mock self.src_msg_runner.instance_fault_create_at_top( self.ctxt, fake_instance_fault) diff --git a/nova/tests/fake_instance.py b/nova/tests/fake_instance.py index e91cf8009c..b1a080269d 100644 --- a/nova/tests/fake_instance.py +++ b/nova/tests/fake_instance.py @@ -17,7 +17,6 @@ import uuid from nova import objects from nova.objects import fields -from nova.objects import instance_fault as inst_fault_obj def fake_db_secgroups(instance, names): @@ -103,6 +102,6 @@ def fake_fault_obj(context, instance_uuid, code=404, } if updates: fault.update(updates) - return inst_fault_obj.InstanceFault._from_db_object(context, - inst_fault_obj.InstanceFault(), - fault) + return objects.InstanceFault._from_db_object(context, + objects.InstanceFault(), + fault) diff --git a/nova/tests/scheduler/test_filter_scheduler.py b/nova/tests/scheduler/test_filter_scheduler.py index a38138c31c..7551a7eec5 100644 --- a/nova/tests/scheduler/test_filter_scheduler.py +++ b/nova/tests/scheduler/test_filter_scheduler.py @@ -27,7 +27,7 @@ from nova.compute import vm_states from nova import context from nova import db from nova import exception -from nova.objects import instance_group as instance_group_obj +from nova import objects from nova.pci import pci_request from nova.scheduler import driver from nova.scheduler import filter_scheduler @@ -375,7 +375,7 @@ class FilterSchedulerTestCase(test_scheduler.SchedulerTestCase): instance = fake_instance.fake_instance_obj(self.context, params={'host': 'hostA'}) - group = instance_group_obj.InstanceGroup() + group = objects.InstanceGroup() group.name = 'pele' group.uuid = str(uuid.uuid4()) group.members = [instance.uuid] @@ -393,10 +393,9 @@ class FilterSchedulerTestCase(test_scheduler.SchedulerTestCase): } with contextlib.nested( - mock.patch.object(instance_group_obj.InstanceGroup, func, - return_value=group), - mock.patch.object(instance_group_obj.InstanceGroup, 'get_hosts', - return_value=['hostA']), + mock.patch.object(objects.InstanceGroup, func, return_value=group), + mock.patch.object(objects.InstanceGroup, 'get_hosts', + return_value=['hostA']), ) as (get_group, get_hosts): update_group_hosts = sched._setup_instance_group(self.context, filter_properties) diff --git a/nova/tests/virt/baremetal/test_pxe.py b/nova/tests/virt/baremetal/test_pxe.py index d0c8f52111..73eb764d3a 100644 --- a/nova/tests/virt/baremetal/test_pxe.py +++ b/nova/tests/virt/baremetal/test_pxe.py @@ -25,7 +25,7 @@ from oslo.config import cfg from testtools import matchers from nova import exception -from nova.objects import flavor as flavor_obj +from nova import objects from nova.openstack.common.db import exception as db_exc from nova.tests.image import fake as fake_image from nova.tests import utils @@ -437,15 +437,15 @@ class PXEPublicMethodsTestCase(BareMetalPXETestCase): def test_cache_images(self): self._create_node() - self.mox.StubOutWithMock(flavor_obj.Flavor, 'get_by_id') + self.mox.StubOutWithMock(objects.Flavor, 'get_by_id') self.mox.StubOutWithMock(pxe, "get_tftp_image_info") self.mox.StubOutWithMock(self.driver, "_cache_tftp_images") self.mox.StubOutWithMock(self.driver, "_cache_image") self.mox.StubOutWithMock(self.driver, "_inject_into_image") - flavor_obj.Flavor.get_by_id(self.context, - self.instance['instance_type_id'] - ).AndReturn({}) + objects.Flavor.get_by_id(self.context, + self.instance['instance_type_id'] + ).AndReturn({}) pxe.get_tftp_image_info(self.instance, {}).AndReturn([]) self.driver._cache_tftp_images(self.context, self.instance, []) self.driver._cache_image(self.context, self.instance, []) @@ -501,7 +501,7 @@ class PXEPublicMethodsTestCase(BareMetalPXETestCase): pxe_path = pxe.get_pxe_config_file_path(self.instance) pxe.get_image_file_path(self.instance) - self.mox.StubOutWithMock(flavor_obj.Flavor, 'get_by_id') + self.mox.StubOutWithMock(objects.Flavor, 'get_by_id') self.mox.StubOutWithMock(pxe, 'get_tftp_image_info') self.mox.StubOutWithMock(pxe, 'get_partition_sizes') self.mox.StubOutWithMock(bm_utils, 'random_alnum') @@ -509,9 +509,9 @@ class PXEPublicMethodsTestCase(BareMetalPXETestCase): self.mox.StubOutWithMock(bm_utils, 'write_to_file') self.mox.StubOutWithMock(bm_utils, 'create_link_without_raise') - flavor_obj.Flavor.get_by_id(self.context, - self.instance['instance_type_id'] - ).AndReturn({}) + objects.Flavor.get_by_id(self.context, + self.instance['instance_type_id'] + ).AndReturn({}) pxe.get_tftp_image_info(self.instance, {}).AndReturn(image_info) pxe.get_partition_sizes(self.instance).AndReturn((0, 0, 0)) bm_utils.random_alnum(32).AndReturn('alnum') @@ -533,7 +533,7 @@ class PXEPublicMethodsTestCase(BareMetalPXETestCase): def test_activate_and_deactivate_bootloader(self): self._create_node() - flavor = flavor_obj.Flavor( + flavor = objects.Flavor( context=self.context, extra_specs={ 'baremetal:deploy_kernel_id': 'eeee', @@ -541,13 +541,13 @@ class PXEPublicMethodsTestCase(BareMetalPXETestCase): }) self.instance['uuid'] = 'fake-uuid' - self.mox.StubOutWithMock(flavor_obj.Flavor, 'get_by_id') + self.mox.StubOutWithMock(objects.Flavor, 'get_by_id') self.mox.StubOutWithMock(bm_utils, 'write_to_file') self.mox.StubOutWithMock(bm_utils, 'create_link_without_raise') self.mox.StubOutWithMock(bm_utils, 'unlink_without_raise') self.mox.StubOutWithMock(bm_utils, 'rmtree_without_raise') - flavor_obj.Flavor.get_by_id( + objects.Flavor.get_by_id( self.context, self.instance['instance_type_id']).AndReturn( flavor) diff --git a/nova/tests/virt/libvirt/test_driver.py b/nova/tests/virt/libvirt/test_driver.py index 970a3b69a6..79908847ee 100644 --- a/nova/tests/virt/libvirt/test_driver.py +++ b/nova/tests/virt/libvirt/test_driver.py @@ -46,9 +46,6 @@ from nova import db from nova import exception from nova.network import model as network_model from nova import objects -from nova.objects import flavor as flavor_obj -from nova.objects import pci_device as pci_device_obj -from nova.objects import service as service_obj from nova.openstack.common import fileutils from nova.openstack.common import importutils from nova.openstack.common import jsonutils @@ -634,7 +631,7 @@ class LibvirtConnTestCase(test.TestCase, def create_instance_obj(self, context, **params): default_params = self.test_instance - default_params['pci_devices'] = pci_device_obj.PciDeviceList() + default_params['pci_devices'] = objects.PciDeviceList() default_params.update(params) instance = objects.Instance(context, **params) flavor = flavors.get_default_flavor() @@ -902,7 +899,7 @@ class LibvirtConnTestCase(test.TestCase, mock.patch.object(conn, "_connect", return_value=self.conn), mock.patch.object(self.conn, "registerCloseCallback", side_effect=set_close_callback), - mock.patch.object(service_obj.Service, "get_by_compute_host", + mock.patch.object(objects.Service, "get_by_compute_host", return_value=service_mock)): # verify that the driver registers for the close callback @@ -929,7 +926,7 @@ class LibvirtConnTestCase(test.TestCase, mock.patch.object(conn, "_connect", return_value=self.conn), mock.patch.object(self.conn, "registerCloseCallback", side_effect=TypeError('dd')), - mock.patch.object(service_obj.Service, "get_by_compute_host", + mock.patch.object(objects.Service, "get_by_compute_host", return_value=service_mock)): connection = conn._get_connection() @@ -947,7 +944,7 @@ class LibvirtConnTestCase(test.TestCase, mock.patch.object(conn, "_connect", return_value=self.conn), mock.patch.object(self.conn, "registerCloseCallback", side_effect=AttributeError('dd')), - mock.patch.object(service_obj.Service, "get_by_compute_host", + mock.patch.object(objects.Service, "get_by_compute_host", return_value=service_mock)): connection = conn._get_connection() @@ -1552,9 +1549,8 @@ class LibvirtConnTestCase(test.TestCase, conn = libvirt_driver.LibvirtDriver(fake.FakeVirtAPI(), True) - fake_flavor = flavor_obj.Flavor.get_by_id( - self.context, - self.test_instance['instance_type_id']) + fake_flavor = objects.Flavor.get_by_id( + self.context, self.test_instance['instance_type_id']) fake_flavor.extra_specs = {'hw_watchdog_action': 'none'} instance_ref = db.instance_create(self.context, self.test_instance) @@ -1562,7 +1558,7 @@ class LibvirtConnTestCase(test.TestCase, disk_info = blockinfo.get_disk_info(CONF.libvirt.virt_type, instance_ref) - with mock.patch.object(flavor_obj.Flavor, 'get_by_id', + with mock.patch.object(objects.Flavor, 'get_by_id', return_value=fake_flavor): cfg = conn._get_guest_config(instance_ref, [], {}, disk_info) @@ -1591,9 +1587,8 @@ class LibvirtConnTestCase(test.TestCase, conn = libvirt_driver.LibvirtDriver(fake.FakeVirtAPI(), True) - fake_flavor = flavor_obj.Flavor.get_by_id( - self.context, - self.test_instance['instance_type_id']) + fake_flavor = objects.Flavor.get_by_id( + self.context, self.test_instance['instance_type_id']) fake_flavor.extra_specs = {'hw_watchdog_action': 'none'} instance_ref = db.instance_create(self.context, self.test_instance) @@ -1603,7 +1598,7 @@ class LibvirtConnTestCase(test.TestCase, image_meta = {"properties": {"hw_watchdog_action": "pause"}} - with mock.patch.object(flavor_obj.Flavor, 'get_by_id', + with mock.patch.object(objects.Flavor, 'get_by_id', return_value=fake_flavor): cfg = conn._get_guest_config(instance_ref, [], image_meta, disk_info) @@ -1713,7 +1708,7 @@ class LibvirtConnTestCase(test.TestCase, agent_enabled=True, group='spice') - instance_type = flavor_obj.Flavor.get_by_id(self.context, 5) + instance_type = objects.Flavor.get_by_id(self.context, 5) instance_type.extra_specs = {'hw_video:ram_max_mb': "100"} conn = libvirt_driver.LibvirtDriver(fake.FakeVirtAPI(), True) instance_ref = db.instance_create(self.context, self.test_instance) @@ -1722,7 +1717,7 @@ class LibvirtConnTestCase(test.TestCase, instance_ref) image_meta = {"properties": {"hw_video_model": "qxl", "hw_video_ram": "64"}} - with mock.patch.object(flavor_obj.Flavor, 'get_by_id', + with mock.patch.object(objects.Flavor, 'get_by_id', return_value=instance_type): cfg = conn._get_guest_config(instance_ref, [], image_meta, disk_info) @@ -1772,7 +1767,7 @@ class LibvirtConnTestCase(test.TestCase, agent_enabled=True, group='spice') - instance_type = flavor_obj.Flavor.get_by_id(self.context, 5) + instance_type = objects.Flavor.get_by_id(self.context, 5) instance_type.extra_specs = {'hw_video:ram_max_mb': "50"} conn = libvirt_driver.LibvirtDriver(fake.FakeVirtAPI(), True) instance_ref = db.instance_create(self.context, self.test_instance) @@ -1781,7 +1776,7 @@ class LibvirtConnTestCase(test.TestCase, instance_ref) image_meta = {"properties": {"hw_video_model": "qxl", "hw_video_ram": "64"}} - with mock.patch.object(flavor_obj.Flavor, 'get_by_id', + with mock.patch.object(objects.Flavor, 'get_by_id', return_value=instance_type): self.assertRaises(exception.RequestedVRamTooHigh, conn._get_guest_config, @@ -1824,7 +1819,7 @@ class LibvirtConnTestCase(test.TestCase, use_usb_tablet=False, group='libvirt') - fake_flavor = flavor_obj.Flavor.get_by_id( + fake_flavor = objects.Flavor.get_by_id( self.context, self.test_instance['instance_type_id']) fake_flavor.extra_specs = {'hw_rng:allowed': 'True'} @@ -1835,7 +1830,7 @@ class LibvirtConnTestCase(test.TestCase, instance_ref) image_meta = {"properties": {"hw_rng_model": "virtio"}} - with mock.patch.object(flavor_obj.Flavor, 'get_by_id', + with mock.patch.object(objects.Flavor, 'get_by_id', return_value=fake_flavor): cfg = conn._get_guest_config(instance_ref, [], image_meta, disk_info) @@ -1892,7 +1887,7 @@ class LibvirtConnTestCase(test.TestCase, use_usb_tablet=False, group='libvirt') - fake_flavor = flavor_obj.Flavor.get_by_id( + fake_flavor = objects.Flavor.get_by_id( self.context, self.test_instance['instance_type_id']) fake_flavor.extra_specs = {'hw_rng:allowed': 'True', @@ -1905,7 +1900,7 @@ class LibvirtConnTestCase(test.TestCase, instance_ref) image_meta = {"properties": {"hw_rng_model": "virtio"}} - with mock.patch.object(flavor_obj.Flavor, 'get_by_id', + with mock.patch.object(objects.Flavor, 'get_by_id', return_value=fake_flavor): cfg = conn._get_guest_config(instance_ref, [], image_meta, disk_info) @@ -1935,7 +1930,7 @@ class LibvirtConnTestCase(test.TestCase, rng_dev_path='/dev/hw_rng', group='libvirt') - fake_flavor = flavor_obj.Flavor.get_by_id( + fake_flavor = objects.Flavor.get_by_id( self.context, self.test_instance['instance_type_id']) fake_flavor.extra_specs = {'hw_rng:allowed': 'True'} @@ -1946,7 +1941,7 @@ class LibvirtConnTestCase(test.TestCase, instance_ref) image_meta = {"properties": {"hw_rng_model": "virtio"}} - with contextlib.nested(mock.patch.object(flavor_obj.Flavor, + with contextlib.nested(mock.patch.object(objects.Flavor, 'get_by_id', return_value=fake_flavor), mock.patch('nova.virt.libvirt.driver.os.path.exists', @@ -1979,7 +1974,7 @@ class LibvirtConnTestCase(test.TestCase, rng_dev_path='/dev/hw_rng', group='libvirt') - fake_flavor = flavor_obj.Flavor.get_by_id( + fake_flavor = objects.Flavor.get_by_id( self.context, self.test_instance['instance_type_id']) fake_flavor.extra_specs = {'hw_rng:allowed': 'True'} @@ -1990,7 +1985,7 @@ class LibvirtConnTestCase(test.TestCase, instance_ref) image_meta = {"properties": {"hw_rng_model": "virtio"}} - with contextlib.nested(mock.patch.object(flavor_obj.Flavor, + with contextlib.nested(mock.patch.object(objects.Flavor, 'get_by_id', return_value=fake_flavor), mock.patch('nova.virt.libvirt.driver.os.path.exists', @@ -2006,7 +2001,7 @@ class LibvirtConnTestCase(test.TestCase, conn = libvirt_driver.LibvirtDriver(fake.FakeVirtAPI(), True) - fake_flavor = flavor_obj.Flavor.get_by_id( + fake_flavor = objects.flavor.Flavor.get_by_id( self.context, self.test_instance['instance_type_id']) fake_flavor.extra_specs = {'quota:cpu_shares': '10000', @@ -2017,7 +2012,7 @@ class LibvirtConnTestCase(test.TestCase, disk_info = blockinfo.get_disk_info(CONF.libvirt.virt_type, instance_ref) - with mock.patch.object(flavor_obj.Flavor, 'get_by_id', + with mock.patch.object(objects.flavor.Flavor, 'get_by_id', return_value=fake_flavor): cfg = conn._get_guest_config(instance_ref, [], {}, disk_info) @@ -2029,7 +2024,7 @@ class LibvirtConnTestCase(test.TestCase, conn = libvirt_driver.LibvirtDriver(fake.FakeVirtAPI(), True) - fake_flavor = flavor_obj.Flavor.get_by_id( + fake_flavor = objects.flavor.Flavor.get_by_id( self.context, self.test_instance['instance_type_id']) fake_flavor.extra_specs = {'quota:cpu_shares': 'fishfood', @@ -2040,7 +2035,7 @@ class LibvirtConnTestCase(test.TestCase, disk_info = blockinfo.get_disk_info(CONF.libvirt.virt_type, instance_ref) - with mock.patch.object(flavor_obj.Flavor, 'get_by_id', + with mock.patch.object(objects.flavor.Flavor, 'get_by_id', return_value=fake_flavor): self.assertRaises(ValueError, conn._get_guest_config, @@ -2557,7 +2552,7 @@ class LibvirtConnTestCase(test.TestCase, self.assertEqual(conf.cpu.threads, 1) def test_get_guest_cpu_topology(self): - fake_flavor = flavor_obj.Flavor.get_by_id( + fake_flavor = objects.flavor.Flavor.get_by_id( self.context, self.test_instance['instance_type_id']) fake_flavor.vcpus = 8 @@ -2568,7 +2563,7 @@ class LibvirtConnTestCase(test.TestCase, disk_info = blockinfo.get_disk_info(CONF.libvirt.virt_type, instance_ref) - with mock.patch.object(flavor_obj.Flavor, 'get_by_id', + with mock.patch.object(objects.flavor.Flavor, 'get_by_id', return_value=fake_flavor): conf = conn._get_guest_config(instance_ref, _fake_network_info(self.stubs, 1), @@ -5624,7 +5619,7 @@ class LibvirtConnTestCase(test.TestCase, with contextlib.nested( mock.patch.object(libvirt, 'openAuth', return_value=mock.MagicMock()), - mock.patch.object(service_obj.Service, "get_by_compute_host", + mock.patch.object(objects.Service, "get_by_compute_host", return_value=service_mock)): conn.get_num_instances() @@ -7540,7 +7535,7 @@ Active: 8381604 kB else: raise ValueError("Unhandled method %" % method_name) - fake_flavor = flavor_obj.Flavor.get_by_id( + fake_flavor = objects.Flavor.get_by_id( self.context, test_instance['instance_type_id']) expected = conn.vif_driver.get_config(test_instance, network_info[0], fake_image_meta, @@ -7548,7 +7543,7 @@ Active: 8381604 kB self.mox.StubOutWithMock(conn.vif_driver, 'get_config') conn.vif_driver.get_config(test_instance, network_info[0], fake_image_meta, - mox.IsA(flavor_obj.Flavor)).\ + mox.IsA(objects.Flavor)).\ AndReturn(expected) self.mox.ReplayAll() @@ -9963,7 +9958,7 @@ class LibvirtDriverTestCase(test.TestCase): self.libvirtconnection.firewall_driver.setup_basic_filtering( instance, [network_info[0]]) - fake_flavor = flavor_obj.Flavor.get_by_id( + fake_flavor = objects.Flavor.get_by_id( self.context, instance['instance_type_id']) if method == 'attach_interface': fake_image_meta = {'id': instance['image_ref']} @@ -9977,7 +9972,7 @@ class LibvirtDriverTestCase(test.TestCase): self.libvirtconnection.vif_driver.get_config( instance, network_info[0], fake_image_meta, - mox.IsA(flavor_obj.Flavor)).AndReturn(expected) + mox.IsA(objects.Flavor)).AndReturn(expected) domain.info().AndReturn([power_state]) if method == 'attach_interface': domain.attachDeviceFlags(expected.to_xml(), expected_flags) diff --git a/nova/tests/virt/test_virt_drivers.py b/nova/tests/virt/test_virt_drivers.py index 947a67a7e0..ff51d4c346 100644 --- a/nova/tests/virt/test_virt_drivers.py +++ b/nova/tests/virt/test_virt_drivers.py @@ -23,6 +23,7 @@ import six from nova.compute import manager from nova import exception +from nova import objects from nova.openstack.common import importutils from nova.openstack.common import jsonutils from nova.openstack.common import log as logging @@ -784,8 +785,7 @@ class LibvirtConnTestCase(_VirtDriverTestCase, test.TestCase): # Previous status of the service: disabled: False service_mock.configure_mock(disabled_reason='None', disabled=False) - from nova.objects import service as service_obj - with mock.patch.object(service_obj.Service, "get_by_compute_host", + with mock.patch.object(objects.Service, "get_by_compute_host", return_value=service_mock): self.connection._set_host_enabled(False, 'ERROR!') self.assertTrue(service_mock.disabled) @@ -798,8 +798,7 @@ class LibvirtConnTestCase(_VirtDriverTestCase, test.TestCase): # Previous status of the service: disabled: True, 'AUTO: ERROR' service_mock.configure_mock(disabled_reason='AUTO: ERROR', disabled=True) - from nova.objects import service as service_obj - with mock.patch.object(service_obj.Service, "get_by_compute_host", + with mock.patch.object(objects.Service, "get_by_compute_host", return_value=service_mock): self.connection._set_host_enabled(True) self.assertFalse(service_mock.disabled) @@ -812,8 +811,7 @@ class LibvirtConnTestCase(_VirtDriverTestCase, test.TestCase): # Previous status of the service: disabled: True, 'Manually disabled' service_mock.configure_mock(disabled_reason='Manually disabled', disabled=True) - from nova.objects import service as service_obj - with mock.patch.object(service_obj.Service, "get_by_compute_host", + with mock.patch.object(objects.Service, "get_by_compute_host", return_value=service_mock): self.connection._set_host_enabled(True) self.assertTrue(service_mock.disabled) @@ -826,8 +824,7 @@ class LibvirtConnTestCase(_VirtDriverTestCase, test.TestCase): # Previous status of the service: disabled: True, 'Manually disabled' service_mock.configure_mock(disabled_reason='Manually disabled', disabled=True) - from nova.objects import service as service_obj - with mock.patch.object(service_obj.Service, "get_by_compute_host", + with mock.patch.object(objects.Service, "get_by_compute_host", return_value=service_mock): self.connection._set_host_enabled(False, 'ERROR!') self.assertTrue(service_mock.disabled) diff --git a/nova/virt/baremetal/pxe.py b/nova/virt/baremetal/pxe.py index 5b57aa2928..72d5a02169 100644 --- a/nova/virt/baremetal/pxe.py +++ b/nova/virt/baremetal/pxe.py @@ -27,7 +27,7 @@ from oslo.config import cfg from nova.compute import flavors from nova import exception from nova.i18n import _ -from nova.objects import flavor as flavor_obj +from nova import objects from nova.openstack.common.db import exception as db_exc from nova.openstack.common import fileutils from nova.openstack.common import log as logging @@ -334,8 +334,8 @@ class PXE(base.NodeDriver): def cache_images(self, context, node, instance, admin_password, image_meta, injected_files, network_info): """Prepare all the images for this instance.""" - flavor = flavor_obj.Flavor.get_by_id(context, - instance['instance_type_id']) + flavor = objects.Flavor.get_by_id(context, + instance['instance_type_id']) tftp_image_info = get_tftp_image_info(instance, flavor) self._cache_tftp_images(context, instance, tftp_image_info) @@ -379,8 +379,8 @@ class PXE(base.NodeDriver): ./pxelinux.cfg/ {mac} -> ../{uuid}/config """ - flavor = flavor_obj.Flavor.get_by_id(context, - instance['instance_type_id']) + flavor = objects.Flavor.get_by_id(context, + instance['instance_type_id']) image_info = get_tftp_image_info(instance, flavor) (root_mb, swap_mb, ephemeral_mb) = get_partition_sizes(instance) pxe_config_file_path = get_pxe_config_file_path(instance) diff --git a/nova/virt/firewall.py b/nova/virt/firewall.py index ed3ff026af..452edbf05a 100644 --- a/nova/virt/firewall.py +++ b/nova/virt/firewall.py @@ -23,8 +23,6 @@ from nova.i18n import _ from nova.i18n import _LI from nova.network import linux_net from nova import objects -from nova.objects import security_group as security_group_obj -from nova.objects import security_group_rule as security_group_rule_obj from nova.openstack.common import importutils from nova.openstack.common import log as logging from nova import utils @@ -357,13 +355,13 @@ class IptablesFirewallDriver(FirewallDriver): # Allow RA responses self._do_ra_rules(ipv6_rules, network_info) - security_groups = security_group_obj.SecurityGroupList.get_by_instance( + security_groups = objects.SecurityGroupList.get_by_instance( ctxt, instance) # then, security group chains and rules for security_group in security_groups: - rules_cls = security_group_rule_obj.SecurityGroupRuleList - rules = rules_cls.get_by_security_group(ctxt, security_group) + rules = objects.SecurityGroupRuleList.get_by_security_group( + ctxt, security_group) for rule in rules: if not rule['cidr']: diff --git a/nova/virt/libvirt/driver.py b/nova/virt/libvirt/driver.py index f4f89bb2da..5da030bbaa 100644 --- a/nova/virt/libvirt/driver.py +++ b/nova/virt/libvirt/driver.py @@ -62,8 +62,6 @@ from nova.i18n import _LI from nova.i18n import _LW from nova import image from nova import objects -from nova.objects import flavor as flavor_obj -from nova.objects import service as service_obj from nova.openstack.common import excutils from nova.openstack.common import fileutils from nova.openstack.common import importutils @@ -1422,7 +1420,7 @@ class LibvirtDriver(driver.ComputeDriver): def attach_interface(self, instance, image_meta, vif): virt_dom = self._lookup_by_name(instance['name']) - flavor = flavor_obj.Flavor.get_by_id( + flavor = objects.Flavor.get_by_id( nova_context.get_admin_context(read_deleted='yes'), instance['instance_type_id']) self.vif_driver.plug(instance, vif) @@ -1443,7 +1441,7 @@ class LibvirtDriver(driver.ComputeDriver): def detach_interface(self, instance, vif): virt_dom = self._lookup_by_name(instance['name']) - flavor = flavor_obj.Flavor.get_by_id( + flavor = objects.Flavor.get_by_id( nova_context.get_admin_context(read_deleted='yes'), instance['instance_type_id']) cfg = self.vif_driver.get_config(instance, vif, None, flavor) @@ -2881,7 +2879,7 @@ class LibvirtDriver(driver.ComputeDriver): ctx = nova_context.get_admin_context() try: - service = service_obj.Service.get_by_compute_host(ctx, CONF.host) + service = objects.Service.get_by_compute_host(ctx, CONF.host) if service.disabled != disable_service: # Note(jang): this is a quick fix to stop operator- @@ -3183,7 +3181,7 @@ class LibvirtDriver(driver.ComputeDriver): 'kernel_id' if a kernel is needed for the rescue image. """ - flavor = flavor_obj.Flavor.get_by_id( + flavor = objects.Flavor.get_by_id( nova_context.get_admin_context(read_deleted='yes'), instance['instance_type_id']) inst_path = libvirt_utils.get_instance_path(instance) diff --git a/nova/virt/xenapi/host.py b/nova/virt/xenapi/host.py index c842df8a7b..e66c0eda90 100644 --- a/nova/virt/xenapi/host.py +++ b/nova/virt/xenapi/host.py @@ -27,7 +27,6 @@ from nova import context from nova import exception from nova.i18n import _ from nova import objects -from nova.objects import service as service_obj from nova.openstack.common import jsonutils from nova.openstack.common import log as logging from nova.pci import pci_whitelist @@ -120,8 +119,8 @@ class Host(object): # Since capabilities are gone, use service table to disable a node # in scheduler cntxt = context.get_admin_context() - service = service_obj.Service.get_by_args(cntxt, CONF.host, - 'nova-compute') + service = objects.Service.get_by_args(cntxt, CONF.host, + 'nova-compute') service.disabled = not enabled service.disabled_reason = 'set by xenapi host_state' service.save() |