summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorDavanum Srinivas <davanum@gmail.com>2015-02-20 09:24:15 -0500
committerDavanum Srinivas <davanum@gmail.com>2015-02-22 07:56:09 -0500
commit69d0c643610dadca24fd146b6fdcfd82362fb431 (patch)
treeb57599a6a6be9560838ea70ba3cf1283b1e87dae
parent50e0da9a44c6a75b97cdbd2e8d4a9c83da34c376 (diff)
downloadnova-69d0c643610dadca24fd146b6fdcfd82362fb431.tar.gz
switch LOG.audit to LOG.info
oslo.log does not have AUDIT level. This was a conscious decision documented in blueprint: https://blueprints.launchpad.net/oslo.log/+spec/remove-context-adapter which was implemented in: I10240f8af6c42508124659b3ed62c5ab93552953 So to prep for switching to oslo.log, this changeset removes usage of LOG.audit and replaces it with LOG.info Change-Id: I9cb1293529b2079b8a4778e27d60c6c760dfb622
-rw-r--r--nova/api/ec2/__init__.py3
-rw-r--r--nova/api/ec2/cloud.py36
-rw-r--r--nova/api/ec2/ec2utils.py3
-rw-r--r--nova/api/openstack/compute/contrib/assisted_volume_snapshots.py6
-rw-r--r--nova/api/openstack/compute/contrib/attach_interfaces.py5
-rw-r--r--nova/api/openstack/compute/contrib/hosts.py7
-rw-r--r--nova/api/openstack/compute/contrib/server_external_events.py5
-rw-r--r--nova/api/openstack/compute/contrib/volumes.py15
-rw-r--r--nova/api/openstack/compute/plugins/v3/assisted_volume_snapshots.py6
-rw-r--r--nova/api/openstack/compute/plugins/v3/hosts.py7
-rw-r--r--nova/api/openstack/compute/plugins/v3/server_external_events.py5
-rw-r--r--nova/compute/api.py4
-rw-r--r--nova/compute/claims.py11
-rw-r--r--nova/compute/manager.py41
-rw-r--r--nova/compute/resource_tracker.py16
-rw-r--r--nova/consoleauth/manager.py6
-rw-r--r--nova/service.py4
-rw-r--r--nova/tests/unit/test_hacking.py2
-rw-r--r--nova/virt/block_device.py3
-rw-r--r--nova/vnc/xvp_proxy.py19
20 files changed, 108 insertions, 96 deletions
diff --git a/nova/api/ec2/__init__.py b/nova/api/ec2/__init__.py
index 83999987a2..94a669f9bd 100644
--- a/nova/api/ec2/__init__.py
+++ b/nova/api/ec2/__init__.py
@@ -40,6 +40,7 @@ from nova import context
from nova import exception
from nova.i18n import _
from nova.i18n import _LE
+from nova.i18n import _LI
from nova.i18n import _LW
from nova.openstack.common import log as logging
from nova.openstack.common import memorycache
@@ -461,7 +462,7 @@ class Authorizer(wsgi.Middleware):
if self._matches_any_role(context, allowed_roles):
return self.application
else:
- LOG.audit(_('Unauthorized request for controller=%(controller)s '
+ LOG.info(_LI('Unauthorized request for controller=%(controller)s '
'and action=%(action)s'),
{'controller': controller, 'action': action},
context=context)
diff --git a/nova/api/ec2/cloud.py b/nova/api/ec2/cloud.py
index 723c6972e5..0c2899ca4b 100644
--- a/nova/api/ec2/cloud.py
+++ b/nova/api/ec2/cloud.py
@@ -39,6 +39,7 @@ from nova.compute import api as compute_api
from nova.compute import vm_states
from nova import exception
from nova.i18n import _
+from nova.i18n import _LI
from nova.i18n import _LW
from nova.image import s3
from nova import network
@@ -403,7 +404,7 @@ class CloudController(object):
def create_snapshot(self, context, volume_id, **kwargs):
validate_volume_id(volume_id)
- LOG.audit(_("Create snapshot of volume %s"), volume_id,
+ LOG.info(_LI("Create snapshot of volume %s"), volume_id,
context=context)
volume_id = ec2utils.ec2_vol_id_to_uuid(volume_id)
args = (context, volume_id, kwargs.get('name'),
@@ -446,7 +447,7 @@ class CloudController(object):
return {'keySet': result}
def create_key_pair(self, context, key_name, **kwargs):
- LOG.audit(_("Create key pair %s"), key_name, context=context)
+ LOG.info(_LI("Create key pair %s"), key_name, context=context)
keypair, private_key = self.keypair_api.create_key_pair(
context, context.user_id, key_name)
@@ -458,7 +459,7 @@ class CloudController(object):
def import_key_pair(self, context, key_name, public_key_material,
**kwargs):
- LOG.audit(_("Import key %s"), key_name, context=context)
+ LOG.info(_LI("Import key %s"), key_name, context=context)
public_key = base64.b64decode(public_key_material)
@@ -471,7 +472,7 @@ class CloudController(object):
'keyFingerprint': keypair['fingerprint']}
def delete_key_pair(self, context, key_name, **kwargs):
- LOG.audit(_("Delete key pair %s"), key_name, context=context)
+ LOG.info(_LI("Delete key pair %s"), key_name, context=context)
try:
self.keypair_api.delete_key_pair(context, context.user_id,
key_name)
@@ -774,7 +775,7 @@ class CloudController(object):
"passwordData": output}
def get_console_output(self, context, instance_id, **kwargs):
- LOG.audit(_("Get console output for instance %s"), instance_id,
+ LOG.info(_LI("Get console output for instance %s"), instance_id,
context=context)
# instance_id may be passed in as a list of instances
if isinstance(instance_id, list):
@@ -847,11 +848,11 @@ class CloudController(object):
if snapshot_ec2id is not None:
snapshot_id = ec2utils.ec2_snap_id_to_uuid(kwargs['snapshot_id'])
snapshot = self.volume_api.get_snapshot(context, snapshot_id)
- LOG.audit(_("Create volume from snapshot %s"), snapshot_ec2id,
+ LOG.info(_LI("Create volume from snapshot %s"), snapshot_ec2id,
context=context)
else:
snapshot = None
- LOG.audit(_("Create volume of %s GB"),
+ LOG.info(_LI("Create volume of %s GB"),
kwargs.get('size'),
context=context)
@@ -891,7 +892,7 @@ class CloudController(object):
instance_uuid = ec2utils.ec2_inst_id_to_uuid(context, instance_id)
instance = self.compute_api.get(context, instance_uuid,
want_objects=True)
- LOG.audit(_('Attach volume %(volume_id)s to instance %(instance_id)s '
+ LOG.info(_LI('Attach volume %(volume_id)s to instance %(instance_id)s '
'at %(device)s'),
{'volume_id': volume_id,
'instance_id': instance_id,
@@ -921,7 +922,7 @@ class CloudController(object):
def detach_volume(self, context, volume_id, **kwargs):
validate_volume_id(volume_id)
volume_id = ec2utils.ec2_vol_id_to_uuid(volume_id)
- LOG.audit(_("Detach volume %s"), volume_id, context=context)
+ LOG.info(_LI("Detach volume %s"), volume_id, context=context)
volume = self.volume_api.get(context, volume_id)
instance = self._get_instance_from_volume(context, volume)
@@ -1288,17 +1289,17 @@ class CloudController(object):
return address
def allocate_address(self, context, **kwargs):
- LOG.audit(_("Allocate address"), context=context)
+ LOG.info(_LI("Allocate address"), context=context)
public_ip = self.network_api.allocate_floating_ip(context)
return {'publicIp': public_ip}
def release_address(self, context, public_ip, **kwargs):
- LOG.audit(_('Release address %s'), public_ip, context=context)
+ LOG.info(_LI('Release address %s'), public_ip, context=context)
self.network_api.release_floating_ip(context, address=public_ip)
return {'return': "true"}
def associate_address(self, context, instance_id, public_ip, **kwargs):
- LOG.audit(_("Associate address %(public_ip)s to instance "
+ LOG.info(_LI("Associate address %(public_ip)s to instance "
"%(instance_id)s"),
{'public_ip': public_ip, 'instance_id': instance_id},
context=context)
@@ -1331,7 +1332,8 @@ class CloudController(object):
if instance_id:
instance = self.compute_api.get(context, instance_id,
want_objects=True)
- LOG.audit(_("Disassociate address %s"), public_ip, context=context)
+ LOG.info(_LI("Disassociate address %s"),
+ public_ip, context=context)
self.network_api.disassociate_floating_ip(context, instance,
address=public_ip)
else:
@@ -1483,7 +1485,7 @@ class CloudController(object):
def reboot_instances(self, context, instance_id, **kwargs):
"""instance_id is a list of instance ids."""
instances = self._ec2_ids_to_instances(context, instance_id)
- LOG.audit(_("Reboot instance %r"), instance_id, context=context)
+ LOG.info(_LI("Reboot instance %r"), instance_id, context=context)
for instance in instances:
self.compute_api.reboot(context, instance, 'HARD')
return True
@@ -1598,7 +1600,7 @@ class CloudController(object):
return {'imagesSet': images}
def deregister_image(self, context, image_id, **kwargs):
- LOG.audit(_("De-registering image %s"), image_id, context=context)
+ LOG.info(_LI("De-registering image %s"), image_id, context=context)
image = self._get_image(context, image_id)
internal_id = image['id']
self.image_service.delete(context, internal_id)
@@ -1634,7 +1636,7 @@ class CloudController(object):
metadata['properties']['block_device_mapping'] = mappings
image_id = self._register_image(context, metadata)
- LOG.audit(_('Registered image %(image_location)s with id '
+ LOG.info(_LI('Registered image %(image_location)s with id '
'%(image_id)s'),
{'image_location': image_location, 'image_id': image_id},
context=context)
@@ -1703,7 +1705,7 @@ class CloudController(object):
if operation_type not in ['add', 'remove']:
msg = _('operation_type must be add or remove')
raise exception.InvalidParameterValue(message=msg)
- LOG.audit(_("Updating image %s publicity"), image_id, context=context)
+ LOG.info(_LI("Updating image %s publicity"), image_id, context=context)
try:
image = self._get_image(context, image_id)
diff --git a/nova/api/ec2/ec2utils.py b/nova/api/ec2/ec2utils.py
index 0acfb95bd7..a2c7a52c4e 100644
--- a/nova/api/ec2/ec2utils.py
+++ b/nova/api/ec2/ec2utils.py
@@ -22,6 +22,7 @@ from oslo_utils import timeutils
from nova import context
from nova import exception
from nova.i18n import _
+from nova.i18n import _LI
from nova.network import model as network_model
from nova import objects
from nova.objects import base as obj_base
@@ -288,7 +289,7 @@ def is_ec2_timestamp_expired(request, expires=None):
timeutils.is_newer_than(query_time, expires)
return False
except ValueError:
- LOG.audit(_("Timestamp is invalid."))
+ LOG.info(_LI("Timestamp is invalid."))
return True
diff --git a/nova/api/openstack/compute/contrib/assisted_volume_snapshots.py b/nova/api/openstack/compute/contrib/assisted_volume_snapshots.py
index c5bb8238ce..2df4cd4e2a 100644
--- a/nova/api/openstack/compute/contrib/assisted_volume_snapshots.py
+++ b/nova/api/openstack/compute/contrib/assisted_volume_snapshots.py
@@ -20,7 +20,7 @@ from nova.api.openstack import extensions
from nova.api.openstack import wsgi
from nova import compute
from nova import exception
-from nova.i18n import _
+from nova.i18n import _LI
from nova.openstack.common import log as logging
@@ -50,7 +50,7 @@ class AssistedVolumeSnapshotsController(wsgi.Controller):
except KeyError:
raise webob.exc.HTTPBadRequest()
- LOG.audit(_("Create assisted snapshot from volume %s"), volume_id,
+ LOG.info(_LI("Create assisted snapshot from volume %s"), volume_id,
context=context)
return self.compute_api.volume_snapshot_create(context, volume_id,
@@ -61,7 +61,7 @@ class AssistedVolumeSnapshotsController(wsgi.Controller):
context = req.environ['nova.context']
authorize(context, action='delete')
- LOG.audit(_("Delete snapshot with id: %s"), id, context=context)
+ LOG.info(_LI("Delete snapshot with id: %s"), id, context=context)
delete_metadata = {}
delete_metadata.update(req.GET)
diff --git a/nova/api/openstack/compute/contrib/attach_interfaces.py b/nova/api/openstack/compute/contrib/attach_interfaces.py
index 6f3a12db37..b16d830b2a 100644
--- a/nova/api/openstack/compute/contrib/attach_interfaces.py
+++ b/nova/api/openstack/compute/contrib/attach_interfaces.py
@@ -25,6 +25,7 @@ from nova.api.openstack import extensions
from nova import compute
from nova import exception
from nova.i18n import _
+from nova.i18n import _LI
from nova import network
from nova.openstack.common import log as logging
@@ -117,7 +118,7 @@ class InterfaceAttachmentController(object):
instance = common.get_instance(self.compute_api,
context, server_id,
want_objects=True)
- LOG.audit(_("Attach interface"), instance=instance)
+ LOG.info(_LI("Attach interface"), instance=instance)
vif = self.compute_api.attach_interface(context,
instance, network_id, port_id, req_ip)
except (exception.PortNotFound,
@@ -149,7 +150,7 @@ class InterfaceAttachmentController(object):
instance = common.get_instance(self.compute_api,
context, server_id,
want_objects=True)
- LOG.audit(_("Detach interface %s"), port_id, instance=instance)
+ LOG.info(_LI("Detach interface %s"), port_id, instance=instance)
try:
self.compute_api.detach_interface(context,
instance, port_id=port_id)
diff --git a/nova/api/openstack/compute/contrib/hosts.py b/nova/api/openstack/compute/contrib/hosts.py
index 9fbefd7309..46a17728d5 100644
--- a/nova/api/openstack/compute/contrib/hosts.py
+++ b/nova/api/openstack/compute/contrib/hosts.py
@@ -21,6 +21,7 @@ from nova.api.openstack import extensions
from nova import compute
from nova import exception
from nova.i18n import _
+from nova.i18n import _LI
from nova import objects
from nova.openstack.common import log as logging
@@ -142,7 +143,7 @@ class HostController(object):
"""Start/Stop host maintenance window. On start, it triggers
guest VMs evacuation.
"""
- LOG.audit(_("Putting host %(host_name)s in maintenance mode "
+ LOG.info(_LI("Putting host %(host_name)s in maintenance mode "
"%(mode)s."),
{'host_name': host_name, 'mode': mode})
try:
@@ -165,9 +166,9 @@ class HostController(object):
on the host
"""
if enabled:
- LOG.audit(_("Enabling host %s.") % host_name)
+ LOG.info(_LI("Enabling host %s.") % host_name)
else:
- LOG.audit(_("Disabling host %s.") % host_name)
+ LOG.info(_LI("Disabling host %s.") % host_name)
try:
result = self.api.set_host_enabled(context, host_name=host_name,
enabled=enabled)
diff --git a/nova/api/openstack/compute/contrib/server_external_events.py b/nova/api/openstack/compute/contrib/server_external_events.py
index c3be2266c4..d6b4183942 100644
--- a/nova/api/openstack/compute/contrib/server_external_events.py
+++ b/nova/api/openstack/compute/contrib/server_external_events.py
@@ -19,6 +19,7 @@ from nova.api.openstack import wsgi
from nova import compute
from nova import exception
from nova.i18n import _
+from nova.i18n import _LI
from nova import objects
from nova.objects import external_event as external_event_obj
from nova.openstack.common import log as logging
@@ -93,8 +94,8 @@ class ServerExternalEventsController(wsgi.Controller):
if instance.host:
accepted_events.append(event)
accepted_instances.add(instance)
- LOG.audit(_('Creating event %(name)s:%(tag)s for instance '
- '%(instance_uuid)s'),
+ LOG.info(_LI('Creating event %(name)s:%(tag)s for '
+ 'instance %(instance_uuid)s'),
dict(event.iteritems()))
# NOTE: as the event is processed asynchronously verify
# whether 202 is a more suitable response code than 200
diff --git a/nova/api/openstack/compute/contrib/volumes.py b/nova/api/openstack/compute/contrib/volumes.py
index 4806aaf940..5c7a734df0 100644
--- a/nova/api/openstack/compute/contrib/volumes.py
+++ b/nova/api/openstack/compute/contrib/volumes.py
@@ -25,6 +25,7 @@ from nova.api.openstack import wsgi
from nova import compute
from nova import exception
from nova.i18n import _
+from nova.i18n import _LI
from nova import objects
from nova.openstack.common import log as logging
from nova.openstack.common import uuidutils
@@ -73,7 +74,7 @@ def _translate_volume_summary_view(context, vol):
d['volumeType'] = vol['volume_type_id']
d['snapshotId'] = vol['snapshot_id']
- LOG.audit(_("vol=%s"), vol, context=context)
+ LOG.info(_LI("vol=%s"), vol, context=context)
if vol.get('volume_metadata'):
d['metadata'] = vol.get('volume_metadata')
@@ -107,7 +108,7 @@ class VolumeController(wsgi.Controller):
context = req.environ['nova.context']
authorize(context)
- LOG.audit(_("Delete volume with id: %s"), id, context=context)
+ LOG.info(_LI("Delete volume with id: %s"), id, context=context)
try:
self.volume_api.delete(context, id)
@@ -159,7 +160,7 @@ class VolumeController(wsgi.Controller):
if size is None and snapshot is not None:
size = snapshot['volume_size']
- LOG.audit(_("Create volume of %s GB"), size, context=context)
+ LOG.info(_LI("Create volume of %s GB"), size, context=context)
availability_zone = vol.get('availability_zone', None)
@@ -292,7 +293,7 @@ class VolumeAttachmentController(wsgi.Controller):
self._validate_volume_id(volume_id)
- LOG.audit(_("Attach volume %(volume_id)s to instance %(server_id)s "
+ LOG.info(_LI("Attach volume %(volume_id)s to instance %(server_id)s "
"at %(device)s"),
{'volume_id': volume_id,
'device': device,
@@ -391,7 +392,7 @@ class VolumeAttachmentController(wsgi.Controller):
authorize_attach(context, action='delete')
volume_id = id
- LOG.audit(_("Detach volume %s"), volume_id, context=context)
+ LOG.info(_LI("Detach volume %s"), volume_id, context=context)
instance = common.get_instance(self.compute_api, context, server_id,
want_objects=True)
@@ -502,7 +503,7 @@ class SnapshotController(wsgi.Controller):
context = req.environ['nova.context']
authorize(context)
- LOG.audit(_("Delete snapshot with id: %s"), id, context=context)
+ LOG.info(_LI("Delete snapshot with id: %s"), id, context=context)
try:
self.volume_api.delete_snapshot(context, id)
@@ -540,7 +541,7 @@ class SnapshotController(wsgi.Controller):
snapshot = body['snapshot']
volume_id = snapshot['volume_id']
- LOG.audit(_("Create snapshot from volume %s"), volume_id,
+ LOG.info(_LI("Create snapshot from volume %s"), volume_id,
context=context)
force = snapshot.get('force', False)
diff --git a/nova/api/openstack/compute/plugins/v3/assisted_volume_snapshots.py b/nova/api/openstack/compute/plugins/v3/assisted_volume_snapshots.py
index f9cae99599..3f8deff7ac 100644
--- a/nova/api/openstack/compute/plugins/v3/assisted_volume_snapshots.py
+++ b/nova/api/openstack/compute/plugins/v3/assisted_volume_snapshots.py
@@ -26,7 +26,7 @@ from nova.api.openstack import wsgi
from nova.api import validation
from nova import compute
from nova import exception
-from nova.i18n import _
+from nova.i18n import _LI
from nova.openstack.common import log as logging
@@ -54,7 +54,7 @@ class AssistedVolumeSnapshotsController(wsgi.Controller):
create_info = snapshot['create_info']
volume_id = snapshot['volume_id']
- LOG.audit(_("Create assisted snapshot from volume %s"), volume_id,
+ LOG.info(_LI("Create assisted snapshot from volume %s"), volume_id,
context=context)
try:
return self.compute_api.volume_snapshot_create(context, volume_id,
@@ -70,7 +70,7 @@ class AssistedVolumeSnapshotsController(wsgi.Controller):
context = req.environ['nova.context']
authorize(context, action='delete')
- LOG.audit(_("Delete snapshot with id: %s"), id, context=context)
+ LOG.info(_LI("Delete snapshot with id: %s"), id, context=context)
delete_metadata = {}
delete_metadata.update(req.GET)
diff --git a/nova/api/openstack/compute/plugins/v3/hosts.py b/nova/api/openstack/compute/plugins/v3/hosts.py
index 2a8cf28db2..e64335dcce 100644
--- a/nova/api/openstack/compute/plugins/v3/hosts.py
+++ b/nova/api/openstack/compute/plugins/v3/hosts.py
@@ -24,6 +24,7 @@ from nova.api import validation
from nova import compute
from nova import exception
from nova.i18n import _
+from nova.i18n import _LI
from nova import objects
from nova.openstack.common import log as logging
@@ -133,7 +134,7 @@ class HostController(wsgi.Controller):
"""Start/Stop host maintenance window. On start, it triggers
guest VMs evacuation.
"""
- LOG.audit(_("Putting host %(host_name)s in maintenance mode "
+ LOG.info(_LI("Putting host %(host_name)s in maintenance mode "
"%(mode)s."),
{'host_name': host_name, 'mode': mode})
try:
@@ -155,9 +156,9 @@ class HostController(wsgi.Controller):
on the host.
"""
if enabled:
- LOG.audit(_("Enabling host %s."), host_name)
+ LOG.info(_LI("Enabling host %s."), host_name)
else:
- LOG.audit(_("Disabling host %s."), host_name)
+ LOG.info(_LI("Disabling host %s."), host_name)
try:
result = self.api.set_host_enabled(context, host_name=host_name,
enabled=enabled)
diff --git a/nova/api/openstack/compute/plugins/v3/server_external_events.py b/nova/api/openstack/compute/plugins/v3/server_external_events.py
index 5890b5e63d..a3ebecdbff 100644
--- a/nova/api/openstack/compute/plugins/v3/server_external_events.py
+++ b/nova/api/openstack/compute/plugins/v3/server_external_events.py
@@ -21,6 +21,7 @@ from nova.api import validation
from nova import compute
from nova import exception
from nova.i18n import _
+from nova.i18n import _LI
from nova import objects
from nova.openstack.common import log as logging
@@ -83,8 +84,8 @@ class ServerExternalEventsController(wsgi.Controller):
if instance.host:
accepted_events.append(event)
accepted_instances.add(instance)
- LOG.audit(_('Creating event %(name)s:%(tag)s for instance '
- '%(instance_uuid)s'),
+ LOG.info(_LI('Creating event %(name)s:%(tag)s for '
+ 'instance %(instance_uuid)s'),
dict(event.iteritems()))
# NOTE: as the event is processed asynchronously verify
# whether 202 is a more suitable response code than 200
diff --git a/nova/compute/api.py b/nova/compute/api.py
index fac9164144..9b5f823b16 100644
--- a/nova/compute/api.py
+++ b/nova/compute/api.py
@@ -3867,7 +3867,7 @@ class SecurityGroupAPI(base.Base, security_group_base.SecurityGroupBase):
msg = _("Quota exceeded, too many security groups.")
self.raise_over_quota(msg)
- LOG.audit(_("Create Security Group %s"), name, context=context)
+ LOG.info(_LI("Create Security Group %s"), name, context=context)
try:
self.ensure_default(context)
@@ -3972,7 +3972,7 @@ class SecurityGroupAPI(base.Base, security_group_base.SecurityGroupBase):
LOG.exception(_LE("Failed to update usages deallocating "
"security group"))
- LOG.audit(_("Delete security group %s"), security_group['name'],
+ LOG.info(_LI("Delete security group %s"), security_group['name'],
context=context)
self.db.security_group_destroy(context, security_group['id'])
diff --git a/nova/compute/claims.py b/nova/compute/claims.py
index e0e31caa4a..5d64305030 100644
--- a/nova/compute/claims.py
+++ b/nova/compute/claims.py
@@ -22,6 +22,7 @@ from oslo_serialization import jsonutils
from nova import context
from nova import exception
from nova.i18n import _
+from nova.i18n import _LI
from nova import objects
from nova.objects import base as obj_base
from nova.openstack.common import log as logging
@@ -166,7 +167,7 @@ class Claim(NopClaim):
raise exception.ComputeResourcesUnavailable(reason=
"; ".join(reasons))
- LOG.audit(_('Claim successful'), instance=self.instance)
+ LOG.info(_LI('Claim successful'), instance=self.instance)
def _test_memory(self, resources, limit):
type_ = _("memory")
@@ -235,22 +236,22 @@ class Claim(NopClaim):
"""Test if the given type of resource needed for a claim can be safely
allocated.
"""
- LOG.audit(_('Total %(type)s: %(total)d %(unit)s, used: %(used).02f '
+ LOG.info(_LI('Total %(type)s: %(total)d %(unit)s, used: %(used).02f '
'%(unit)s'),
{'type': type_, 'total': total, 'unit': unit, 'used': used},
instance=self.instance)
if limit is None:
# treat resource as unlimited:
- LOG.audit(_('%(type)s limit not specified, defaulting to '
+ LOG.info(_LI('%(type)s limit not specified, defaulting to '
'unlimited'), {'type': type_}, instance=self.instance)
return
free = limit - used
# Oversubscribed resource policy info:
- LOG.audit(_('%(type)s limit: %(limit).02f %(unit)s, free: %(free).02f '
- '%(unit)s'),
+ LOG.info(_LI('%(type)s limit: %(limit).02f %(unit)s, '
+ 'free: %(free).02f %(unit)s'),
{'type': type_, 'limit': limit, 'free': free, 'unit': unit},
instance=self.instance)
diff --git a/nova/compute/manager.py b/nova/compute/manager.py
index e5b4fe793d..9a9874f1f0 100644
--- a/nova/compute/manager.py
+++ b/nova/compute/manager.py
@@ -1650,7 +1650,7 @@ class ComputeManager(manager.Manager):
def _start_building(self, context, instance):
"""Save the host and launched_on fields and log appropriately."""
- LOG.audit(_('Starting instance...'), context=context,
+ LOG.info(_LI('Starting instance...'), context=context,
instance=instance)
self._instance_update(context, instance.uuid,
vm_state=vm_states.BUILDING,
@@ -2055,7 +2055,7 @@ class ComputeManager(manager.Manager):
node=None, limits=None):
try:
- LOG.audit(_('Starting instance...'), context=context,
+ LOG.info(_LI('Starting instance...'), context=context,
instance=instance)
instance.vm_state = vm_states.BUILDING
instance.task_state = None
@@ -2427,7 +2427,8 @@ class ComputeManager(manager.Manager):
trying to teardown networking
"""
context = context.elevated()
- LOG.audit(_('%(action_str)s instance') % {'action_str': 'Terminating'},
+ LOG.info(_LI('%(action_str)s instance') %
+ {'action_str': 'Terminating'},
context=context, instance=instance)
if notify:
@@ -2794,7 +2795,7 @@ class ComputeManager(manager.Manager):
orig_vm_state = instance.vm_state
with self._error_out_instance_on_exception(context, instance):
- LOG.audit(_("Rebuilding instance"), context=context,
+ LOG.info(_LI("Rebuilding instance"), context=context,
instance=instance)
if recreate:
@@ -2961,7 +2962,7 @@ class ComputeManager(manager.Manager):
task_states.REBOOT_PENDING_HARD,
task_states.REBOOT_STARTED_HARD)
context = context.elevated()
- LOG.audit(_("Rebooting instance"), context=context, instance=instance)
+ LOG.info(_LI("Rebooting instance"), context=context, instance=instance)
block_device_info = self._get_instance_block_device_info(context,
instance)
@@ -3099,7 +3100,7 @@ class ComputeManager(manager.Manager):
try:
instance.save()
- LOG.audit(_('instance snapshotting'), context=context,
+ LOG.info(_LI('instance snapshotting'), context=context,
instance=instance)
if instance.power_state != power_state.RUNNING:
@@ -3234,7 +3235,7 @@ class ComputeManager(manager.Manager):
try:
self.driver.set_admin_password(instance, new_pass)
- LOG.audit(_("Root password set"), instance=instance)
+ LOG.info(_LI("Root password set"), instance=instance)
instance.task_state = None
instance.save(
expected_task_state=task_states.UPDATING_PASSWORD)
@@ -3282,7 +3283,7 @@ class ComputeManager(manager.Manager):
{'current_state': current_power_state,
'expected_state': expected_state},
instance=instance)
- LOG.audit(_('injecting file to %s'), path,
+ LOG.info(_LI('injecting file to %s'), path,
instance=instance)
self.driver.inject_file(instance, path, file_contents)
@@ -3320,7 +3321,7 @@ class ComputeManager(manager.Manager):
def rescue_instance(self, context, instance, rescue_password,
rescue_image_ref=None, clean_shutdown=True):
context = context.elevated()
- LOG.audit(_('Rescuing'), context=context, instance=instance)
+ LOG.info(_LI('Rescuing'), context=context, instance=instance)
admin_password = (rescue_password if rescue_password else
utils.generate_password())
@@ -3369,7 +3370,7 @@ class ComputeManager(manager.Manager):
@wrap_instance_fault
def unrescue_instance(self, context, instance):
context = context.elevated()
- LOG.audit(_('Unrescuing'), context=context, instance=instance)
+ LOG.info(_LI('Unrescuing'), context=context, instance=instance)
network_info = self._get_instance_nw_info(context, instance)
self._notify_about_instance_usage(context, instance,
@@ -3703,7 +3704,7 @@ class ComputeManager(manager.Manager):
rt = self._get_resource_tracker(node)
with rt.resize_claim(context, instance, instance_type,
image_meta=image, limits=limits) as claim:
- LOG.audit(_('Migrating'), context=context, instance=instance)
+ LOG.info(_LI('Migrating'), context=context, instance=instance)
self.compute_rpcapi.resize_instance(
context, instance, claim.migration, image,
instance_type, quotas.reservations,
@@ -4042,7 +4043,7 @@ class ComputeManager(manager.Manager):
def pause_instance(self, context, instance):
"""Pause an instance on this host."""
context = context.elevated()
- LOG.audit(_('Pausing'), context=context, instance=instance)
+ LOG.info(_LI('Pausing'), context=context, instance=instance)
self._notify_about_instance_usage(context, instance, 'pause.start')
self.driver.pause(instance)
instance.power_state = self._get_power_state(context, instance)
@@ -4058,7 +4059,7 @@ class ComputeManager(manager.Manager):
def unpause_instance(self, context, instance):
"""Unpause a paused instance on this host."""
context = context.elevated()
- LOG.audit(_('Unpausing'), context=context, instance=instance)
+ LOG.info(_LI('Unpausing'), context=context, instance=instance)
self._notify_about_instance_usage(context, instance, 'unpause.start')
self.driver.unpause(instance)
instance.power_state = self._get_power_state(context, instance)
@@ -4096,7 +4097,7 @@ class ComputeManager(manager.Manager):
"""Retrieve diagnostics for an instance on this host."""
current_power_state = self._get_power_state(context, instance)
if current_power_state == power_state.RUNNING:
- LOG.audit(_("Retrieving diagnostics"), context=context,
+ LOG.info(_LI("Retrieving diagnostics"), context=context,
instance=instance)
return self.driver.get_diagnostics(instance)
else:
@@ -4113,7 +4114,7 @@ class ComputeManager(manager.Manager):
"""Retrieve diagnostics for an instance on this host."""
current_power_state = self._get_power_state(context, instance)
if current_power_state == power_state.RUNNING:
- LOG.audit(_("Retrieving diagnostics"), context=context,
+ LOG.info(_LI("Retrieving diagnostics"), context=context,
instance=instance)
diags = self.driver.get_instance_diagnostics(instance)
return diags.serialize()
@@ -4152,7 +4153,7 @@ class ComputeManager(manager.Manager):
def resume_instance(self, context, instance):
"""Resume the given suspended instance."""
context = context.elevated()
- LOG.audit(_('Resuming'), context=context, instance=instance)
+ LOG.info(_LI('Resuming'), context=context, instance=instance)
self._notify_about_instance_usage(context, instance, 'resume.start')
network_info = self._get_instance_nw_info(context, instance)
@@ -4388,7 +4389,7 @@ class ComputeManager(manager.Manager):
def get_console_output(self, context, instance, tail_length):
"""Send the console output for the given instance."""
context = context.elevated()
- LOG.audit(_("Get console output"), context=context,
+ LOG.info(_LI("Get console output"), context=context,
instance=instance)
output = self.driver.get_console_output(context, instance)
@@ -4632,7 +4633,7 @@ class ComputeManager(manager.Manager):
def _attach_volume(self, context, instance, bdm):
context = context.elevated()
- LOG.audit(_('Attaching volume %(volume_id)s to %(mountpoint)s'),
+ LOG.info(_LI('Attaching volume %(volume_id)s to %(mountpoint)s'),
{'volume_id': bdm.volume_id,
'mountpoint': bdm['mount_device']},
context=context, instance=instance)
@@ -4657,7 +4658,7 @@ class ComputeManager(manager.Manager):
mp = bdm.device_name
volume_id = bdm.volume_id
- LOG.audit(_('Detach volume %(volume_id)s from mountpoint %(mp)s'),
+ LOG.info(_LI('Detach volume %(volume_id)s from mountpoint %(mp)s'),
{'volume_id': volume_id, 'mp': mp},
context=context, instance=instance)
@@ -6068,7 +6069,7 @@ class ComputeManager(manager.Manager):
for cn in compute_nodes_in_db:
if cn.hypervisor_hostname not in nodenames:
- LOG.audit(_("Deleting orphan compute node %s") % cn.id)
+ LOG.info(_LI("Deleting orphan compute node %s") % cn.id)
cn.destroy()
self._resource_tracker_dict = new_resource_tracker_dict
diff --git a/nova/compute/resource_tracker.py b/nova/compute/resource_tracker.py
index 5cbdd9e00f..ae20412840 100644
--- a/nova/compute/resource_tracker.py
+++ b/nova/compute/resource_tracker.py
@@ -311,12 +311,12 @@ class ResourceTracker(object):
declared a need for resources, but not necessarily retrieved them from
the hypervisor layer yet.
"""
- LOG.audit(_("Auditing locally available compute resources"))
+ LOG.info(_LI("Auditing locally available compute resources"))
resources = self.driver.get_available_resource(self.nodename)
if not resources:
# The virt driver does not support this function
- LOG.audit(_("Virt driver does not support "
+ LOG.info(_LI("Virt driver does not support "
"'get_available_resource' Compute tracking is disabled."))
self.compute_node = None
return
@@ -510,25 +510,25 @@ class ResourceTracker(object):
including instance calculations and in-progress resource claims. These
values will be exposed via the compute node table to the scheduler.
"""
- LOG.audit(_("Total physical ram (MB): %(pram)s, "
+ LOG.info(_LI("Total physical ram (MB): %(pram)s, "
"total allocated virtual ram (MB): %(vram)s"),
{'pram': resources['memory_mb'],
'vram': resources['memory_mb_used']})
- LOG.audit(_("Total physical disk (GB): %(pdisk)s, "
+ LOG.info(_LI("Total physical disk (GB): %(pdisk)s, "
"total allocated virtual disk (GB): %(vdisk)s"),
{'pdisk': resources['local_gb'],
'vdisk': resources['local_gb_used']})
vcpus = resources['vcpus']
if vcpus:
- LOG.audit(_("Total usable vcpus: %(tcpu)s, "
+ LOG.info(_LI("Total usable vcpus: %(tcpu)s, "
"total allocated vcpus: %(ucpu)s"),
{'tcpu': vcpus, 'ucpu': resources['vcpus_used']})
else:
- LOG.audit(_("Free VCPU information unavailable"))
+ LOG.info(_LI("Free VCPU information unavailable"))
if 'pci_stats' in resources:
- LOG.audit(_("PCI stats: %s"), resources['pci_stats'])
+ LOG.info(_LI("PCI stats: %s"), resources['pci_stats'])
def _resource_change(self, resources):
"""Check to see if any resouces have changed."""
@@ -594,7 +594,7 @@ class ResourceTracker(object):
represent an incoming or outbound migration.
"""
uuid = migration['instance_uuid']
- LOG.audit(_("Updating from migration %s") % uuid)
+ LOG.info(_LI("Updating from migration %s") % uuid)
incoming = (migration['dest_compute'] == self.host and
migration['dest_node'] == self.nodename)
diff --git a/nova/consoleauth/manager.py b/nova/consoleauth/manager.py
index dcbbe6b0a7..dd3efb2c36 100644
--- a/nova/consoleauth/manager.py
+++ b/nova/consoleauth/manager.py
@@ -24,7 +24,7 @@ from oslo_serialization import jsonutils
from nova.cells import rpcapi as cells_rpcapi
from nova.compute import rpcapi as compute_rpcapi
-from nova.i18n import _, _LW
+from nova.i18n import _LI, _LW
from nova import manager
from nova import objects
from nova.openstack.common import log as logging
@@ -96,7 +96,7 @@ class ConsoleAuthManager(manager.Manager):
"into memcached"),
{'instance_uuid': instance_uuid})
- LOG.audit(_("Received Token: %(token)s, %(token_dict)s"),
+ LOG.info(_LI("Received Token: %(token)s, %(token_dict)s"),
{'token': token, 'token_dict': token_dict})
def _validate_token(self, context, token):
@@ -121,7 +121,7 @@ class ConsoleAuthManager(manager.Manager):
def check_token(self, context, token):
token_str = self.mc.get(token.encode('UTF-8'))
token_valid = (token_str is not None)
- LOG.audit(_("Checking Token: %(token)s, %(token_valid)s"),
+ LOG.info(_LI("Checking Token: %(token)s, %(token_valid)s"),
{'token': token, 'token_valid': token_valid})
if token_valid:
token = jsonutils.loads(token_str)
diff --git a/nova/service.py b/nova/service.py
index dcc3323422..ad3731a397 100644
--- a/nova/service.py
+++ b/nova/service.py
@@ -31,7 +31,7 @@ from nova import conductor
from nova import context
from nova import debugger
from nova import exception
-from nova.i18n import _, _LE, _LW
+from nova.i18n import _, _LE, _LI, _LW
from nova.objects import base as objects_base
from nova.openstack.common import log as logging
from nova.openstack.common import service
@@ -158,7 +158,7 @@ class Service(service.Service):
def start(self):
verstr = version.version_string_with_package()
- LOG.audit(_('Starting %(topic)s node (version %(version)s)'),
+ LOG.info(_LI('Starting %(topic)s node (version %(version)s)'),
{'topic': self.topic, 'version': verstr})
self.basic_config_check()
self.manager.init_host()
diff --git a/nova/tests/unit/test_hacking.py b/nova/tests/unit/test_hacking.py
index b32b2ce7bb..1a8911b9e3 100644
--- a/nova/tests/unit/test_hacking.py
+++ b/nova/tests/unit/test_hacking.py
@@ -252,7 +252,7 @@ class HackingTestCase(test.NoDBTestCase):
'exception']
levels = ['_LI', '_LW', '_LE', '_LC']
debug = "LOG.debug('OK')"
- audit = "LOG.audit(_('OK'))"
+ audit = "LOG.info(_LI('OK'))"
self.assertEqual(
0, len(list(checks.validate_log_translations(debug, debug, 'f'))))
self.assertEqual(
diff --git a/nova/virt/block_device.py b/nova/virt/block_device.py
index 1d23ad078b..db1909888d 100644
--- a/nova/virt/block_device.py
+++ b/nova/virt/block_device.py
@@ -19,7 +19,6 @@ from oslo_serialization import jsonutils
from oslo_utils import excutils
from nova import block_device
-from nova.i18n import _
from nova.i18n import _LE
from nova.i18n import _LI
from nova import objects
@@ -392,7 +391,7 @@ def attach_block_devices(block_device_mapping, *attach_args, **attach_kwargs):
def _log_and_attach(bdm):
context = attach_args[0]
instance = attach_args[1]
- LOG.audit(_('Booting with volume %(volume_id)s at %(mountpoint)s'),
+ LOG.info(_LI('Booting with volume %(volume_id)s at %(mountpoint)s'),
{'volume_id': bdm.volume_id,
'mountpoint': bdm['mount_device']},
context=context, instance=instance)
diff --git a/nova/vnc/xvp_proxy.py b/nova/vnc/xvp_proxy.py
index 5eb9e6d5a5..29de1c7495 100644
--- a/nova/vnc/xvp_proxy.py
+++ b/nova/vnc/xvp_proxy.py
@@ -27,7 +27,7 @@ import webob
from nova.consoleauth import rpcapi as consoleauth_rpcapi
from nova import context
-from nova.i18n import _
+from nova.i18n import _LI
from nova.openstack.common import log as logging
from nova import version
from nova import wsgi
@@ -91,12 +91,13 @@ class XCPVNCProxy(object):
data += b
if data.find("\r\n\r\n") != -1:
if not data.split("\r\n")[0].find("200"):
- LOG.audit(_("Error in handshake format: %s"), data)
+ LOG.info(_LI("Error in handshake format: %s"),
+ data)
return
break
if not b or len(data) > 4096:
- LOG.audit(_("Error in handshake: %s"), data)
+ LOG.info(_LI("Error in handshake: %s"), data)
return
client = req.environ['eventlet.input'].get_socket()
@@ -111,7 +112,7 @@ class XCPVNCProxy(object):
t0.wait()
if not sockets.get('client') or not sockets.get('server'):
- LOG.audit(_("Invalid request: %s"), req)
+ LOG.info(_LI("Invalid request: %s"), req)
start_response('400 Invalid Request',
[('content-type', 'text/html')])
return "Invalid Request"
@@ -131,10 +132,10 @@ class XCPVNCProxy(object):
def __call__(self, environ, start_response):
try:
req = webob.Request(environ)
- LOG.audit(_("Request: %s"), req)
+ LOG.info(_LI("Request: %s"), req)
token = req.params.get('token')
if not token:
- LOG.audit(_("Request made with missing token: %s"), req)
+ LOG.info(_LI("Request made with missing token: %s"), req)
start_response('400 Invalid Request',
[('content-type', 'text/html')])
return "Invalid Request"
@@ -144,14 +145,14 @@ class XCPVNCProxy(object):
connect_info = api.check_token(ctxt, token)
if not connect_info:
- LOG.audit(_("Request made with invalid token: %s"), req)
+ LOG.info(_LI("Request made with invalid token: %s"), req)
start_response('401 Not Authorized',
[('content-type', 'text/html')])
return "Not Authorized"
return self.proxy_connection(req, connect_info, start_response)
except Exception as e:
- LOG.audit(_("Unexpected error: %s"), e)
+ LOG.info(_LI("Unexpected error: %s"), e)
class SafeHttpProtocol(eventlet.wsgi.HttpProtocol):
@@ -171,7 +172,7 @@ class SafeHttpProtocol(eventlet.wsgi.HttpProtocol):
def get_wsgi_server():
- LOG.audit(_("Starting nova-xvpvncproxy node (version %s)"),
+ LOG.info(_LI("Starting nova-xvpvncproxy node (version %s)"),
version.version_string_with_package())
return wsgi.Server("XCP VNC Proxy",