turn job templates in jobs via launch

This commit is contained in:
Chris Meyers
2016-08-30 15:04:21 -04:00
parent 4c876b40e4
commit 2cfdee3b21
8 changed files with 108 additions and 19 deletions

View File

@@ -2172,7 +2172,6 @@ class SystemJobCancelSerializer(SystemJobSerializer):
# TODO: # TODO:
class WorkflowJobSerializer(UnifiedJobSerializer): class WorkflowJobSerializer(UnifiedJobSerializer):
@@ -2182,13 +2181,17 @@ class WorkflowJobSerializer(UnifiedJobSerializer):
def get_related(self, obj): def get_related(self, obj):
res = super(WorkflowJobSerializer, self).get_related(obj) res = super(WorkflowJobSerializer, self).get_related(obj)
if obj.system_job_template: if obj.workflow_job_template:
res['workflow_job_template'] = reverse('api:workflow_job_template_detail', res['workflow_job_template'] = reverse('api:workflow_job_template_detail',
args=(obj.workflow_job_template.pk,)) args=(obj.workflow_job_template.pk,))
# TODO: # TODO:
#res['notifications'] = reverse('api:system_job_notifications_list', args=(obj.pk,)) #res['notifications'] = reverse('api:system_job_notifications_list', args=(obj.pk,))
res['workflow_nodes'] = reverse('api:workflow_job_workflow_nodes_list', args=(obj.pk,))
# TODO: Cancel job
'''
if obj.can_cancel or True: if obj.can_cancel or True:
res['cancel'] = reverse('api:workflow_job_cancel', args=(obj.pk,)) res['cancel'] = reverse('api:workflow_job_cancel', args=(obj.pk,))
'''
return res return res

View File

@@ -267,6 +267,7 @@ workflow_job_template_urls = patterns('awx.api.views',
workflow_job_urls = patterns('awx.api.views', workflow_job_urls = patterns('awx.api.views',
url(r'^$', 'workflow_job_list'), url(r'^$', 'workflow_job_list'),
url(r'^(?P<pk>[0-9]+)/$', 'workflow_job_detail'), url(r'^(?P<pk>[0-9]+)/$', 'workflow_job_detail'),
url(r'^(?P<pk>[0-9]+)/workflow_nodes/$', 'workflow_job_workflow_nodes_list'),
# url(r'^(?P<pk>[0-9]+)/cancel/$', 'workflow_job_cancel'), # url(r'^(?P<pk>[0-9]+)/cancel/$', 'workflow_job_cancel'),
#url(r'^(?P<pk>[0-9]+)/notifications/$', 'workflow_job_notifications_list'), #url(r'^(?P<pk>[0-9]+)/notifications/$', 'workflow_job_notifications_list'),
) )

View File

@@ -1775,7 +1775,7 @@ class EnforceParentRelationshipMixin(object):
# HACK: Make request data mutable. # HACK: Make request data mutable.
if getattr(data, '_mutable', None) is False: if getattr(data, '_mutable', None) is False:
data._mutable = True data._mutable = True
data[self.enforce_parent_relationship] = getattr(self.get_parent_object(), '%s_id' % relationship) data[self.enforce_parent_relationship] = getattr(self.get_parent_object(), '%s_id' % self.enforce_parent_relationship)
return super(EnforceParentRelationshipMixin, self).create(request, *args, **kwargs) return super(EnforceParentRelationshipMixin, self).create(request, *args, **kwargs)
class GroupChildrenList(EnforceParentRelationshipMixin, SubListCreateAttachDetachAPIView): class GroupChildrenList(EnforceParentRelationshipMixin, SubListCreateAttachDetachAPIView):
@@ -2702,7 +2702,7 @@ class WorkflowJobTemplateLaunch(GenericAPIView):
new_job = obj.create_unified_job(**request.data) new_job = obj.create_unified_job(**request.data)
new_job.signal_start(**request.data) new_job.signal_start(**request.data)
data = dict(system_job=new_job.id) data = dict(workflow_job=new_job.id)
return Response(data, status=status.HTTP_201_CREATED) return Response(data, status=status.HTTP_201_CREATED)
# TODO: # TODO:
@@ -2723,6 +2723,7 @@ class WorkflowJobTemplateJobsList(SubListAPIView):
parent_model = WorkflowJobTemplate parent_model = WorkflowJobTemplate
relationship = 'jobs' relationship = 'jobs'
parent_key = 'workflow_job_template' parent_key = 'workflow_job_template'
# TODO: # TODO:
class WorkflowJobList(ListCreateAPIView): class WorkflowJobList(ListCreateAPIView):
@@ -2740,6 +2741,16 @@ class WorkflowJobDetail(RetrieveDestroyAPIView):
model = WorkflowJob model = WorkflowJob
serializer_class = WorkflowJobSerializer serializer_class = WorkflowJobSerializer
class WorkflowJobWorkflowNodesList(SubListAPIView):
model = WorkflowNode
serializer_class = WorkflowNodeListSerializer
always_allow_superuser = True # TODO: RBAC
parent_model = WorkflowJob
relationship = 'workflow_job_nodes'
parent_key = 'job'
class SystemJobTemplateList(ListAPIView): class SystemJobTemplateList(ListAPIView):
model = SystemJobTemplate model = SystemJobTemplate

View File

@@ -137,6 +137,13 @@ class SimpleDAG(object):
leafs.append(n) leafs.append(n)
return leafs return leafs
def get_root_nodes(self):
roots = []
for n in self.nodes:
if len(self.get_dependents(n['node_object'])) < 1:
roots.append(n)
return roots
def get_tasks(): def get_tasks():
"""Fetch all Tower tasks that are relevant to the task management """Fetch all Tower tasks that are relevant to the task management
system. system.
@@ -151,9 +158,11 @@ def get_tasks():
ProjectUpdate.objects.filter(status__in=RELEVANT_JOBS)] ProjectUpdate.objects.filter(status__in=RELEVANT_JOBS)]
graph_system_jobs = [sj for sj in graph_system_jobs = [sj for sj in
SystemJob.objects.filter(status__in=RELEVANT_JOBS)] SystemJob.objects.filter(status__in=RELEVANT_JOBS)]
graph_workflow_jobs = [wf for wf in
WorkflowJob.objects.filter(status__in=RELEVANT_JOBS)]
all_actions = sorted(graph_jobs + graph_ad_hoc_commands + graph_inventory_updates + all_actions = sorted(graph_jobs + graph_ad_hoc_commands + graph_inventory_updates +
graph_project_updates + graph_system_jobs, graph_project_updates + graph_system_jobs +
graph_workflow_jobs,
key=lambda task: task.created) key=lambda task: task.created)
return all_actions return all_actions

View File

@@ -0,0 +1,20 @@
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
dependencies = [
('main', '0033_v301_workflow_create'),
]
operations = [
migrations.AlterField(
model_name='workflownode',
name='job',
field=models.ForeignKey(related_name='workflow_job_nodes', on_delete=django.db.models.deletion.SET_NULL, default=None, blank=True, to='main.UnifiedJob', null=True),
),
]

View File

@@ -8,6 +8,7 @@ from django.core.urlresolvers import reverse
# AWX # AWX
from awx.main.models import UnifiedJobTemplate, UnifiedJob from awx.main.models import UnifiedJobTemplate, UnifiedJob
from awx.main.models.notifications import JobNotificationMixin
from awx.main.models.base import BaseModel, CreatedModifiedModel, VarsDictProperty from awx.main.models.base import BaseModel, CreatedModifiedModel, VarsDictProperty
from awx.main.models.rbac import ( from awx.main.models.rbac import (
ROLE_SINGLETON_SYSTEM_ADMINISTRATOR, ROLE_SINGLETON_SYSTEM_ADMINISTRATOR,
@@ -61,7 +62,7 @@ class WorkflowNode(CreatedModifiedModel):
) )
job = models.ForeignKey( job = models.ForeignKey(
'UnifiedJob', 'UnifiedJob',
related_name='workflow_node', related_name='workflow_job_nodes',
blank=True, blank=True,
null=True, null=True,
default=None, default=None,
@@ -96,7 +97,7 @@ class WorkflowJobTemplate(UnifiedJobTemplate, WorkflowJobOptions):
@classmethod @classmethod
def _get_unified_job_field_names(cls): def _get_unified_job_field_names(cls):
# TODO: ADD LABELS # TODO: ADD LABELS
return ['name', 'description', 'extra_vars', 'workflow_nodes'] return ['name', 'description', 'extra_vars',]
def get_absolute_url(self): def get_absolute_url(self):
return reverse('api:workflow_job_template_detail', args=(self.pk,)) return reverse('api:workflow_job_template_detail', args=(self.pk,))
@@ -109,14 +110,53 @@ class WorkflowJobTemplate(UnifiedJobTemplate, WorkflowJobOptions):
# TODO: Notifications # TODO: Notifications
# TODO: Surveys # TODO: Surveys
def create_job(self, **kwargs): #def create_job(self, **kwargs):
''' # '''
Create a new job based on this template. # Create a new job based on this template.
''' # '''
return self.create_unified_job(**kwargs) # return self.create_unified_job(**kwargs)
# TODO: Delete create_unified_job here and explicitly call create_workflow_job() .. figure out where the call is
def create_unified_job(self, **kwargs):
class WorkflowJob(UnifiedJob, WorkflowJobOptions): #def create_workflow_job(self, **kwargs):
#workflow_job = self.create_unified_job(**kwargs)
workflow_job = super(WorkflowJobTemplate, self).create_unified_job(**kwargs)
workflow_job.inherit_jt_workflow_nodes()
return workflow_job
class WorkflowJobInheritNodesMixin(object):
def _inherit_relationship(self, old_node, new_node, node_ids_map, node_type):
old_related_nodes = getattr(old_node, node_type).all()
new_node_type_mgr = getattr(new_node, node_type)
for old_related_node in old_related_nodes:
new_related_node_id = node_ids_map[old_related_node.id]
new_related_node = WorkflowNode.objects.get(id=new_related_node_id)
new_node_type_mgr.add(new_related_node)
def inherit_jt_workflow_nodes(self):
new_nodes = []
old_nodes = self.workflow_job_template.workflow_nodes.all()
node_ids_map = {}
for old_node in old_nodes:
new_node = WorkflowNode.objects.get(id=old_node.pk)
new_node.job = self
new_node.pk = None
new_node.save()
new_nodes.append(new_node)
node_ids_map[old_node.id] = new_node.id
for index, old_node in enumerate(old_nodes):
new_node = new_nodes[index]
for node_type in ['success_nodes', 'failure_nodes', 'always_nodes']:
self._inherit_relationship(old_node, new_node, node_ids_map, node_type)
class WorkflowJob(UnifiedJob, WorkflowJobOptions, JobNotificationMixin, WorkflowJobInheritNodesMixin):
class Meta: class Meta:
app_label = 'main' app_label = 'main'
@@ -158,3 +198,11 @@ class WorkflowJob(UnifiedJob, WorkflowJobOptions):
def task_impact(self): def task_impact(self):
return 0 return 0
# TODO: workflow job notifications
def get_notification_templates(self):
return []
# TODO: workflow job notifications
def get_notification_friendly_name(self):
return "Workflow Job"

View File

@@ -191,7 +191,6 @@ def notify_task_runner(metadata_dict):
def _send_notification_templates(instance, status_str): def _send_notification_templates(instance, status_str):
if status_str not in ['succeeded', 'failed']: if status_str not in ['succeeded', 'failed']:
raise ValueError("status_str must be either succeeded or failed") raise ValueError("status_str must be either succeeded or failed")
print("Instance has some shit in it %s" % instance)
notification_templates = instance.get_notification_templates() notification_templates = instance.get_notification_templates()
if notification_templates: if notification_templates:
all_notification_templates = set(notification_templates.get('success', []) + notification_templates.get('any', [])) all_notification_templates = set(notification_templates.get('success', []) + notification_templates.get('any', []))
@@ -239,8 +238,6 @@ def handle_work_error(self, task_id, subtasks=None):
instance.socketio_emit_status("failed") instance.socketio_emit_status("failed")
if first_instance: if first_instance:
print("Instance type is %s" % first_instance_type)
print("Instance passing along %s" % first_instance.name)
_send_notification_templates(first_instance, 'failed') _send_notification_templates(first_instance, 'failed')
@task() @task()
@@ -1675,7 +1672,6 @@ class RunWorkflowJob(BaseTask):
status, rc, tb = 'error', None, '' status, rc, tb = 'error', None, ''
output_replacements = [] output_replacements = []
try: try:
self.pre_run_hook(instance, **kwargs)
if instance.cancel_flag: if instance.cancel_flag:
instance = self.update_model(instance.pk, status='canceled') instance = self.update_model(instance.pk, status='canceled')
if instance.status != 'running': if instance.status != 'running':
@@ -1692,8 +1688,8 @@ class RunWorkflowJob(BaseTask):
except Exception: except Exception:
if status != 'canceled': if status != 'canceled':
tb = traceback.format_exc() tb = traceback.format_exc()
status = 'successful'
instance = self.update_model(pk, status=status, result_traceback=tb) instance = self.update_model(pk, status=status, result_traceback=tb)
self.post_run_hook(instance, **kwargs)
instance.socketio_emit_status(status) instance.socketio_emit_status(status)
if status != 'successful' and not hasattr(settings, 'CELERY_UNIT_TEST'): if status != 'successful' and not hasattr(settings, 'CELERY_UNIT_TEST'):
# Raising an exception will mark the job as 'failed' in celery # Raising an exception will mark the job as 'failed' in celery

View File

@@ -15,6 +15,7 @@ services:
# - sync # - sync
volumes: volumes:
- "../:/tower_devel" - "../:/tower_devel"
privileged: true
# Postgres Database Container # Postgres Database Container
postgres: postgres: