From 6d4a2a8f8e94f677a05f0dc591148d3919ed09ad Mon Sep 17 00:00:00 2001 From: Wayne Witzel III Date: Thu, 11 Aug 2016 15:06:07 -0400 Subject: [PATCH] converting from socketio to channels websocket --- awx/main/models/jobs.py | 1 - awx/main/models/unified_jobs.py | 1 - awx/main/scheduler/__init__.py | 6 +++--- 3 files changed, 3 insertions(+), 5 deletions(-) diff --git a/awx/main/models/jobs.py b/awx/main/models/jobs.py index 269c987aeb..c47ddfd59a 100644 --- a/awx/main/models/jobs.py +++ b/awx/main/models/jobs.py @@ -1409,4 +1409,3 @@ class SystemJob(UnifiedJob, SystemJobOptions, JobNotificationMixin): def get_notification_friendly_name(self): return "System Job" - diff --git a/awx/main/models/unified_jobs.py b/awx/main/models/unified_jobs.py index a4ae651eb7..3a30906e02 100644 --- a/awx/main/models/unified_jobs.py +++ b/awx/main/models/unified_jobs.py @@ -932,4 +932,3 @@ class UnifiedJob(PolymorphicModel, PasswordFieldsModel, CommonModelNameNotUnique if settings.BROKER_URL.startswith('amqp://'): self._force_cancel() return self.cancel_flag - diff --git a/awx/main/scheduler/__init__.py b/awx/main/scheduler/__init__.py index 03c87d8ddb..63bf006bbe 100644 --- a/awx/main/scheduler/__init__.py +++ b/awx/main/scheduler/__init__.py @@ -122,7 +122,7 @@ def rebuild_graph(): logger.debug("Active celery tasks: " + str(active_tasks)) for task in list(running_celery_tasks): if (task.celery_task_id not in active_tasks and not hasattr(settings, 'IGNORE_CELERY_INSPECTOR')): - # NOTE: Pull status again and make sure it didn't finish in + # NOTE: Pull status again and make sure it didn't finish in # the meantime? task.status = 'failed' task.job_explanation += ' '.join(( @@ -204,7 +204,7 @@ def process_graph(graph, task_capacity): node_type = graph.get_node_type(node_obj) if node_type == 'job': - # clear dependencies because a job can block (not necessarily + # clear dependencies because a job can block (not necessarily # depend) on other jobs that share the same job template node_dependencies = [] @@ -217,7 +217,7 @@ def process_graph(graph, task_capacity): node_obj.start() spawn_workflow_graph_jobs([node_obj]) return process_graph(graph, task_capacity) - + dependent_nodes = [{'type': graph.get_node_type(node_obj), 'id': node_obj.id}] + \ [{'type': graph.get_node_type(n['node_object']), 'id': n['node_object'].id} for n in node_dependencies]