mirror of
https://github.com/ansible/awx.git
synced 2026-03-06 03:01:06 -03:30
Merge pull request #487 from AlanCoding/E722
flake8: comply with new E722 rule
This commit is contained in:
@@ -3775,7 +3775,7 @@ class ActivityStreamSerializer(BaseSerializer):
|
|||||||
def get_object_association(self, obj):
|
def get_object_association(self, obj):
|
||||||
try:
|
try:
|
||||||
return obj.object_relationship_type.split(".")[-1].split("_")[1]
|
return obj.object_relationship_type.split(".")[-1].split("_")[1]
|
||||||
except:
|
except Exception:
|
||||||
pass
|
pass
|
||||||
return ""
|
return ""
|
||||||
|
|
||||||
|
|||||||
@@ -359,7 +359,7 @@ class ApiV1ConfigView(APIView):
|
|||||||
try:
|
try:
|
||||||
settings.LICENSE = {}
|
settings.LICENSE = {}
|
||||||
return Response(status=status.HTTP_204_NO_CONTENT)
|
return Response(status=status.HTTP_204_NO_CONTENT)
|
||||||
except:
|
except Exception:
|
||||||
# FIX: Log
|
# FIX: Log
|
||||||
return Response({"error": _("Failed to remove license (%s)") % has_error}, status=status.HTTP_400_BAD_REQUEST)
|
return Response({"error": _("Failed to remove license (%s)") % has_error}, status=status.HTTP_400_BAD_REQUEST)
|
||||||
|
|
||||||
@@ -3200,7 +3200,7 @@ class WorkflowJobTemplateNodeDetail(WorkflowsEnforcementMixin, RetrieveUpdateDes
|
|||||||
try:
|
try:
|
||||||
obj = self.get_object()
|
obj = self.get_object()
|
||||||
data.update(obj.char_prompts)
|
data.update(obj.char_prompts)
|
||||||
except:
|
except Exception:
|
||||||
pass
|
pass
|
||||||
return super(WorkflowJobTemplateNodeDetail, self).update_raw_data(data)
|
return super(WorkflowJobTemplateNodeDetail, self).update_raw_data(data)
|
||||||
|
|
||||||
|
|||||||
@@ -83,7 +83,7 @@ class URLField(CharField):
|
|||||||
else:
|
else:
|
||||||
netloc = '{}@{}' % (url_parts.username, netloc)
|
netloc = '{}@{}' % (url_parts.username, netloc)
|
||||||
value = urlparse.urlunsplit([url_parts.scheme, netloc, url_parts.path, url_parts.query, url_parts.fragment])
|
value = urlparse.urlunsplit([url_parts.scheme, netloc, url_parts.path, url_parts.query, url_parts.fragment])
|
||||||
except:
|
except Exception:
|
||||||
raise # If something fails here, just fall through and let the validators check it.
|
raise # If something fails here, just fall through and let the validators check it.
|
||||||
super(URLField, self).run_validators(value)
|
super(URLField, self).run_validators(value)
|
||||||
|
|
||||||
|
|||||||
@@ -159,14 +159,14 @@ class SettingsRegistry(object):
|
|||||||
if category_slug == 'user' and for_user:
|
if category_slug == 'user' and for_user:
|
||||||
try:
|
try:
|
||||||
field_instance.default = original_field_instance.to_representation(getattr(self.settings, setting))
|
field_instance.default = original_field_instance.to_representation(getattr(self.settings, setting))
|
||||||
except:
|
except Exception:
|
||||||
logger.warning('Unable to retrieve default value for user setting "%s".', setting, exc_info=True)
|
logger.warning('Unable to retrieve default value for user setting "%s".', setting, exc_info=True)
|
||||||
elif not field_instance.read_only or field_instance.default is empty or field_instance.defined_in_file:
|
elif not field_instance.read_only or field_instance.default is empty or field_instance.defined_in_file:
|
||||||
try:
|
try:
|
||||||
field_instance.default = original_field_instance.to_representation(self.settings._awx_conf_settings._get_default(setting))
|
field_instance.default = original_field_instance.to_representation(self.settings._awx_conf_settings._get_default(setting))
|
||||||
except AttributeError:
|
except AttributeError:
|
||||||
pass
|
pass
|
||||||
except:
|
except Exception:
|
||||||
logger.warning('Unable to retrieve default value for setting "%s".', setting, exc_info=True)
|
logger.warning('Unable to retrieve default value for setting "%s".', setting, exc_info=True)
|
||||||
|
|
||||||
# `PENDO_TRACKING_STATE` is disabled for the open source awx license
|
# `PENDO_TRACKING_STATE` is disabled for the open source awx license
|
||||||
|
|||||||
@@ -366,7 +366,7 @@ class SettingsWrapper(UserSettingsHolder):
|
|||||||
return internal_value
|
return internal_value
|
||||||
else:
|
else:
|
||||||
return field.run_validation(value)
|
return field.run_validation(value)
|
||||||
except:
|
except Exception:
|
||||||
logger.warning(
|
logger.warning(
|
||||||
'The current value "%r" for setting "%s" is invalid.',
|
'The current value "%r" for setting "%s" is invalid.',
|
||||||
value, name, exc_info=True)
|
value, name, exc_info=True)
|
||||||
|
|||||||
@@ -16,7 +16,7 @@ class argv_placeholder(object):
|
|||||||
def __del__(self):
|
def __del__(self):
|
||||||
try:
|
try:
|
||||||
argv_ready(sys.argv)
|
argv_ready(sys.argv)
|
||||||
except:
|
except Exception:
|
||||||
pass
|
pass
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
@@ -105,7 +105,7 @@ def _load_default_license_from_file():
|
|||||||
license_data = json.load(open(license_file))
|
license_data = json.load(open(license_file))
|
||||||
logger.debug('Read license data from "%s".', license_file)
|
logger.debug('Read license data from "%s".', license_file)
|
||||||
return license_data
|
return license_data
|
||||||
except:
|
except Exception:
|
||||||
logger.warning('Could not read license from "%s".', license_file, exc_info=True)
|
logger.warning('Could not read license from "%s".', license_file, exc_info=True)
|
||||||
return {}
|
return {}
|
||||||
|
|
||||||
|
|||||||
@@ -122,7 +122,7 @@ def run_pexpect(args, cwd, env, logfile,
|
|||||||
if cancelled_callback:
|
if cancelled_callback:
|
||||||
try:
|
try:
|
||||||
canceled = cancelled_callback()
|
canceled = cancelled_callback()
|
||||||
except:
|
except Exception:
|
||||||
logger.exception('Could not check cancel callback - canceling immediately')
|
logger.exception('Could not check cancel callback - canceling immediately')
|
||||||
if isinstance(extra_update_fields, dict):
|
if isinstance(extra_update_fields, dict):
|
||||||
extra_update_fields['job_explanation'] = "System error during job execution, check system logs"
|
extra_update_fields['job_explanation'] = "System error during job execution, check system logs"
|
||||||
|
|||||||
@@ -168,7 +168,7 @@ class AnsibleInventoryLoader(object):
|
|||||||
data = json.loads(stdout)
|
data = json.loads(stdout)
|
||||||
if not isinstance(data, dict):
|
if not isinstance(data, dict):
|
||||||
raise TypeError('Returned JSON must be a dictionary, got %s instead' % str(type(data)))
|
raise TypeError('Returned JSON must be a dictionary, got %s instead' % str(type(data)))
|
||||||
except:
|
except Exception:
|
||||||
logger.error('Failed to load JSON from: %s', stdout)
|
logger.error('Failed to load JSON from: %s', stdout)
|
||||||
raise
|
raise
|
||||||
return data
|
return data
|
||||||
|
|||||||
@@ -105,7 +105,7 @@ class Schedule(CommonModel):
|
|||||||
if not isinstance(extra_data, dict):
|
if not isinstance(extra_data, dict):
|
||||||
try:
|
try:
|
||||||
extra_data = json.loads(self.extra_data)
|
extra_data = json.loads(self.extra_data)
|
||||||
except:
|
except Exception:
|
||||||
raise ValidationError(_("Expected JSON"))
|
raise ValidationError(_("Expected JSON"))
|
||||||
|
|
||||||
if extra_data and 'days' in extra_data:
|
if extra_data and 'days' in extra_data:
|
||||||
|
|||||||
@@ -810,7 +810,7 @@ class UnifiedJob(PolymorphicModel, PasswordFieldsModel, CommonModelNameNotUnique
|
|||||||
def result_stdout_size(self):
|
def result_stdout_size(self):
|
||||||
try:
|
try:
|
||||||
return os.stat(self.result_stdout_file).st_size
|
return os.stat(self.result_stdout_file).st_size
|
||||||
except:
|
except Exception:
|
||||||
return len(self.result_stdout)
|
return len(self.result_stdout)
|
||||||
|
|
||||||
def _result_stdout_raw_limited(self, start_line=0, end_line=None, redact_sensitive=True, escape_ascii=False):
|
def _result_stdout_raw_limited(self, start_line=0, end_line=None, redact_sensitive=True, escape_ascii=False):
|
||||||
@@ -1072,7 +1072,7 @@ class UnifiedJob(PolymorphicModel, PasswordFieldsModel, CommonModelNameNotUnique
|
|||||||
update_fields.append('job_explanation')
|
update_fields.append('job_explanation')
|
||||||
instance.save(update_fields=update_fields)
|
instance.save(update_fields=update_fields)
|
||||||
self.websocket_emit_status("canceled")
|
self.websocket_emit_status("canceled")
|
||||||
except: # FIXME: Log this exception!
|
except Exception: # FIXME: Log this exception!
|
||||||
if settings.DEBUG:
|
if settings.DEBUG:
|
||||||
raise
|
raise
|
||||||
|
|
||||||
|
|||||||
@@ -23,7 +23,7 @@ import yaml
|
|||||||
import fcntl
|
import fcntl
|
||||||
try:
|
try:
|
||||||
import psutil
|
import psutil
|
||||||
except:
|
except Exception:
|
||||||
psutil = None
|
psutil = None
|
||||||
|
|
||||||
# Celery
|
# Celery
|
||||||
@@ -104,7 +104,7 @@ def celery_startup(conf=None, **kwargs):
|
|||||||
from awx.main.signals import disable_activity_stream
|
from awx.main.signals import disable_activity_stream
|
||||||
with disable_activity_stream():
|
with disable_activity_stream():
|
||||||
sch.save()
|
sch.save()
|
||||||
except:
|
except Exception:
|
||||||
logger.exception("Failed to rebuild schedule {}.".format(sch))
|
logger.exception("Failed to rebuild schedule {}.".format(sch))
|
||||||
|
|
||||||
|
|
||||||
@@ -113,7 +113,7 @@ def task_set_logger_pre_run(*args, **kwargs):
|
|||||||
try:
|
try:
|
||||||
cache.close()
|
cache.close()
|
||||||
configure_external_logger(settings, is_startup=False)
|
configure_external_logger(settings, is_startup=False)
|
||||||
except:
|
except Exception:
|
||||||
# General exception because LogErrorsTask not used with celery signals
|
# General exception because LogErrorsTask not used with celery signals
|
||||||
logger.exception('Encountered error on initial log configuration.')
|
logger.exception('Encountered error on initial log configuration.')
|
||||||
|
|
||||||
@@ -126,7 +126,7 @@ def inform_cluster_of_shutdown(*args, **kwargs):
|
|||||||
this_inst.save(update_fields=['capacity', 'modified'])
|
this_inst.save(update_fields=['capacity', 'modified'])
|
||||||
logger.warning('Normal shutdown signal for instance {}, '
|
logger.warning('Normal shutdown signal for instance {}, '
|
||||||
'removed self from capacity pool.'.format(this_inst.hostname))
|
'removed self from capacity pool.'.format(this_inst.hostname))
|
||||||
except:
|
except Exception:
|
||||||
# General exception because LogErrorsTask not used with celery signals
|
# General exception because LogErrorsTask not used with celery signals
|
||||||
logger.exception('Encountered problem with normal shutdown signal.')
|
logger.exception('Encountered problem with normal shutdown signal.')
|
||||||
|
|
||||||
@@ -321,7 +321,7 @@ def _send_notification_templates(instance, status_str):
|
|||||||
raise ValueError(_("status_str must be either succeeded or failed"))
|
raise ValueError(_("status_str must be either succeeded or failed"))
|
||||||
try:
|
try:
|
||||||
notification_templates = instance.get_notification_templates()
|
notification_templates = instance.get_notification_templates()
|
||||||
except:
|
except Exception:
|
||||||
logger.warn("No notification template defined for emitting notification")
|
logger.warn("No notification template defined for emitting notification")
|
||||||
notification_templates = None
|
notification_templates = None
|
||||||
if notification_templates:
|
if notification_templates:
|
||||||
@@ -443,7 +443,7 @@ def delete_inventory(self, inventory_id, user_id):
|
|||||||
else:
|
else:
|
||||||
try:
|
try:
|
||||||
user = User.objects.get(id=user_id)
|
user = User.objects.get(id=user_id)
|
||||||
except:
|
except Exception:
|
||||||
user = None
|
user = None
|
||||||
with ignore_inventory_computed_fields(), ignore_inventory_group_removal(), impersonate(user):
|
with ignore_inventory_computed_fields(), ignore_inventory_group_removal(), impersonate(user):
|
||||||
try:
|
try:
|
||||||
@@ -877,7 +877,7 @@ class BaseTask(LogErrorsTask):
|
|||||||
**extra_update_fields)
|
**extra_update_fields)
|
||||||
try:
|
try:
|
||||||
self.final_run_hook(instance, status, **kwargs)
|
self.final_run_hook(instance, status, **kwargs)
|
||||||
except:
|
except Exception:
|
||||||
logger.exception('%s Final run hook errored.', instance.log_format)
|
logger.exception('%s Final run hook errored.', instance.log_format)
|
||||||
instance.websocket_emit_status(status)
|
instance.websocket_emit_status(status)
|
||||||
if status != 'successful' and not hasattr(settings, 'CELERY_UNIT_TEST'):
|
if status != 'successful' and not hasattr(settings, 'CELERY_UNIT_TEST'):
|
||||||
|
|||||||
@@ -437,7 +437,7 @@ def group_factory(inventory):
|
|||||||
def g(name):
|
def g(name):
|
||||||
try:
|
try:
|
||||||
return Group.objects.get(name=name, inventory=inventory)
|
return Group.objects.get(name=name, inventory=inventory)
|
||||||
except:
|
except Exception:
|
||||||
return Group.objects.create(inventory=inventory, name=name)
|
return Group.objects.create(inventory=inventory, name=name)
|
||||||
return g
|
return g
|
||||||
|
|
||||||
@@ -478,7 +478,7 @@ def inventory_source_factory(inventory_factory):
|
|||||||
source = 'file'
|
source = 'file'
|
||||||
try:
|
try:
|
||||||
return inventory.inventory_sources.get(name=name)
|
return inventory.inventory_sources.get(name=name)
|
||||||
except:
|
except Exception:
|
||||||
return inventory.inventory_sources.create(name=name, source=source)
|
return inventory.inventory_sources.create(name=name, source=source)
|
||||||
return invsrc
|
return invsrc
|
||||||
|
|
||||||
|
|||||||
@@ -160,7 +160,7 @@ def get_ansible_version():
|
|||||||
stdout=subprocess.PIPE)
|
stdout=subprocess.PIPE)
|
||||||
result = proc.communicate()[0]
|
result = proc.communicate()[0]
|
||||||
return result.split('\n')[0].replace('ansible', '').strip()
|
return result.split('\n')[0].replace('ansible', '').strip()
|
||||||
except:
|
except Exception:
|
||||||
return 'unknown'
|
return 'unknown'
|
||||||
|
|
||||||
|
|
||||||
@@ -174,7 +174,7 @@ def get_ssh_version():
|
|||||||
stderr=subprocess.PIPE)
|
stderr=subprocess.PIPE)
|
||||||
result = proc.communicate()[1]
|
result = proc.communicate()[1]
|
||||||
return result.split(" ")[0].split("_")[1]
|
return result.split(" ")[0].split("_")[1]
|
||||||
except:
|
except Exception:
|
||||||
return 'unknown'
|
return 'unknown'
|
||||||
|
|
||||||
|
|
||||||
@@ -186,7 +186,7 @@ def get_awx_version():
|
|||||||
try:
|
try:
|
||||||
import pkg_resources
|
import pkg_resources
|
||||||
return pkg_resources.require('awx')[0].version
|
return pkg_resources.require('awx')[0].version
|
||||||
except:
|
except Exception:
|
||||||
return __version__
|
return __version__
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
@@ -152,7 +152,7 @@ class BaseHandler(logging.Handler):
|
|||||||
return self._format_and_send_record(record)
|
return self._format_and_send_record(record)
|
||||||
except (KeyboardInterrupt, SystemExit):
|
except (KeyboardInterrupt, SystemExit):
|
||||||
raise
|
raise
|
||||||
except:
|
except Exception:
|
||||||
self.handleError(record)
|
self.handleError(record)
|
||||||
|
|
||||||
def _get_host(self, scheme='', hostname_only=False):
|
def _get_host(self, scheme='', hostname_only=False):
|
||||||
|
|||||||
@@ -41,7 +41,7 @@ from ansible import constants as C
|
|||||||
|
|
||||||
try:
|
try:
|
||||||
from ansible.cache.base import BaseCacheModule
|
from ansible.cache.base import BaseCacheModule
|
||||||
except:
|
except Exception:
|
||||||
from ansible.plugins.cache.base import BaseCacheModule
|
from ansible.plugins.cache.base import BaseCacheModule
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
@@ -158,7 +158,7 @@ class CustomPdb(Rdb):
|
|||||||
try:
|
try:
|
||||||
socket.gethostbyname('docker.for.mac.localhost')
|
socket.gethostbyname('docker.for.mac.localhost')
|
||||||
host = 'docker.for.mac.localhost'
|
host = 'docker.for.mac.localhost'
|
||||||
except:
|
except Exception:
|
||||||
host = os.popen('ip route').read().split(' ')[2]
|
host = os.popen('ip route').read().split(' ')[2]
|
||||||
sock, port = Rdb.get_avail_port(self, *args, **kwargs)
|
sock, port = Rdb.get_avail_port(self, *args, **kwargs)
|
||||||
socket.socket(socket.AF_INET, socket.SOCK_DGRAM).sendto(
|
socket.socket(socket.AF_INET, socket.SOCK_DGRAM).sendto(
|
||||||
@@ -217,7 +217,7 @@ def telnet(port):
|
|||||||
|
|
||||||
try:
|
try:
|
||||||
s.connect(('0.0.0.0', port))
|
s.connect(('0.0.0.0', port))
|
||||||
except:
|
except Exception:
|
||||||
print 'unable to connect'
|
print 'unable to connect'
|
||||||
return
|
return
|
||||||
print 'connected to 0.0.0.0:%d' % port
|
print 'connected to 0.0.0.0:%d' % port
|
||||||
|
|||||||
Reference in New Issue
Block a user