mirror of
https://github.com/ansible/awx.git
synced 2026-01-19 21:51:26 -03:30
Previously, in some cases, an InventoryUpdate sourced by an SCM project would still run and be successful even after the project it is sourced from failed to update. This would happen because the InventoryUpdate would revert the project back to its last working revision. This behavior is confusing and inconsistent with how we handle jobs (which just refuse to launch when the project is failed). This change pulls out the logic that the job launch serializer and RunJob#pre_run_hook had implemented (independently) to check if the project is in a failed state, and puts it into a method on the Project model. This is then checked in the project launch serializer as well as the inventory update serializer, along with SourceControlMixin#sync_and_copy as a fallback for things that don't run the serializer validation (such as scheduled jobs and WFJT jobs). Signed-off-by: Rick Elrod <rick@elrod.me>
674 lines
26 KiB
Python
674 lines
26 KiB
Python
# Copyright (c) 2015 Ansible, Inc.
|
|
# All Rights Reserved.
|
|
|
|
# Python
|
|
import datetime
|
|
import os
|
|
import urllib.parse as urlparse
|
|
|
|
# Django
|
|
from django.conf import settings
|
|
from django.db import models
|
|
from django.utils.translation import gettext_lazy as _
|
|
from django.utils.encoding import smart_str
|
|
from django.utils.text import slugify
|
|
from django.core.exceptions import ValidationError
|
|
from django.utils.timezone import now, make_aware, get_default_timezone
|
|
|
|
|
|
# AWX
|
|
from awx.api.versioning import reverse
|
|
from awx.main.models.base import PROJECT_UPDATE_JOB_TYPE_CHOICES, PERM_INVENTORY_DEPLOY
|
|
from awx.main.models.events import ProjectUpdateEvent, UnpartitionedProjectUpdateEvent
|
|
from awx.main.models.notifications import (
|
|
NotificationTemplate,
|
|
JobNotificationMixin,
|
|
)
|
|
from awx.main.models.unified_jobs import (
|
|
UnifiedJob,
|
|
UnifiedJobTemplate,
|
|
)
|
|
from awx.main.models.jobs import Job
|
|
from awx.main.models.mixins import ResourceMixin, TaskManagerProjectUpdateMixin, CustomVirtualEnvMixin, RelatedJobsMixin
|
|
from awx.main.utils import update_scm_url, polymorphic
|
|
from awx.main.utils.ansible import skip_directory, could_be_inventory, could_be_playbook
|
|
from awx.main.utils.execution_environments import get_control_plane_execution_environment
|
|
from awx.main.fields import ImplicitRoleField, JSONBlob
|
|
from awx.main.models.rbac import (
|
|
ROLE_SINGLETON_SYSTEM_ADMINISTRATOR,
|
|
ROLE_SINGLETON_SYSTEM_AUDITOR,
|
|
)
|
|
|
|
__all__ = ['Project', 'ProjectUpdate']
|
|
|
|
|
|
class ProjectOptions(models.Model):
|
|
|
|
SCM_TYPE_CHOICES = [
|
|
('', _('Manual')),
|
|
('git', _('Git')),
|
|
('svn', _('Subversion')),
|
|
('insights', _('Red Hat Insights')),
|
|
('archive', _('Remote Archive')),
|
|
]
|
|
|
|
class Meta:
|
|
abstract = True
|
|
|
|
# Project files must be available on the server in folders directly
|
|
# beneath the path specified by settings.PROJECTS_ROOT. There is no way
|
|
# via the API to upload/update a project or its playbooks; this must be
|
|
# done by other means for now.
|
|
|
|
@classmethod
|
|
def get_local_path_choices(cls):
|
|
if os.path.exists(settings.PROJECTS_ROOT):
|
|
paths = [
|
|
x
|
|
for x in os.listdir(settings.PROJECTS_ROOT)
|
|
if (os.path.isdir(os.path.join(settings.PROJECTS_ROOT, x)) and not x.startswith('.') and not x.startswith('_'))
|
|
]
|
|
qs = Project.objects
|
|
used_paths = qs.values_list('local_path', flat=True)
|
|
return [x for x in paths if x not in used_paths]
|
|
else:
|
|
return []
|
|
|
|
local_path = models.CharField(
|
|
max_length=1024, blank=True, help_text=_('Local path (relative to PROJECTS_ROOT) containing ' 'playbooks and related files for this project.')
|
|
)
|
|
|
|
scm_type = models.CharField(
|
|
max_length=8,
|
|
choices=SCM_TYPE_CHOICES,
|
|
blank=True,
|
|
default='',
|
|
verbose_name=_('SCM Type'),
|
|
help_text=_("Specifies the source control system used to store the project."),
|
|
)
|
|
scm_url = models.CharField(
|
|
max_length=1024,
|
|
blank=True,
|
|
default='',
|
|
verbose_name=_('SCM URL'),
|
|
help_text=_("The location where the project is stored."),
|
|
)
|
|
scm_branch = models.CharField(
|
|
max_length=256,
|
|
blank=True,
|
|
default='',
|
|
verbose_name=_('SCM Branch'),
|
|
help_text=_('Specific branch, tag or commit to checkout.'),
|
|
)
|
|
scm_refspec = models.CharField(
|
|
max_length=1024,
|
|
blank=True,
|
|
default='',
|
|
verbose_name=_('SCM refspec'),
|
|
help_text=_('For git projects, an additional refspec to fetch.'),
|
|
)
|
|
scm_clean = models.BooleanField(
|
|
default=False,
|
|
help_text=_('Discard any local changes before syncing the project.'),
|
|
)
|
|
scm_delete_on_update = models.BooleanField(
|
|
default=False,
|
|
help_text=_('Delete the project before syncing.'),
|
|
)
|
|
scm_track_submodules = models.BooleanField(
|
|
default=False,
|
|
help_text=_('Track submodules latest commits on defined branch.'),
|
|
)
|
|
credential = models.ForeignKey(
|
|
'Credential',
|
|
related_name='%(class)ss',
|
|
blank=True,
|
|
null=True,
|
|
default=None,
|
|
on_delete=models.SET_NULL,
|
|
)
|
|
timeout = models.IntegerField(
|
|
blank=True,
|
|
default=0,
|
|
help_text=_("The amount of time (in seconds) to run before the task is canceled."),
|
|
)
|
|
|
|
def clean_scm_type(self):
|
|
return self.scm_type or ''
|
|
|
|
def clean_scm_url(self):
|
|
if self.scm_type == 'insights':
|
|
self.scm_url = settings.INSIGHTS_URL_BASE
|
|
scm_url = str(self.scm_url or '')
|
|
if not self.scm_type:
|
|
return ''
|
|
try:
|
|
scm_url = update_scm_url(self.scm_type, scm_url, check_special_cases=False)
|
|
except ValueError as e:
|
|
raise ValidationError((e.args or (_('Invalid SCM URL.'),))[0])
|
|
scm_url_parts = urlparse.urlsplit(scm_url)
|
|
if self.scm_type and not any(scm_url_parts):
|
|
raise ValidationError(_('SCM URL is required.'))
|
|
return str(self.scm_url or '')
|
|
|
|
def clean_credential(self):
|
|
if not self.scm_type:
|
|
return None
|
|
cred = self.credential
|
|
if not cred and self.scm_type == 'insights':
|
|
raise ValidationError(_("Insights Credential is required for an Insights Project."))
|
|
elif cred:
|
|
if self.scm_type == 'insights':
|
|
if cred.kind != 'insights':
|
|
raise ValidationError(_("Credential kind must be 'insights'."))
|
|
elif cred.kind != 'scm':
|
|
raise ValidationError(_("Credential kind must be 'scm'."))
|
|
try:
|
|
if self.scm_type == 'insights':
|
|
self.scm_url = settings.INSIGHTS_URL_BASE
|
|
scm_url = update_scm_url(self.scm_type, self.scm_url, check_special_cases=False)
|
|
scm_url_parts = urlparse.urlsplit(scm_url)
|
|
# Prefer the username/password in the URL, if provided.
|
|
scm_username = scm_url_parts.username or cred.get_input('username', default='')
|
|
if scm_url_parts.password or cred.has_input('password'):
|
|
scm_password = '********'
|
|
else:
|
|
scm_password = ''
|
|
try:
|
|
update_scm_url(self.scm_type, self.scm_url, scm_username, scm_password)
|
|
except ValueError as e:
|
|
raise ValidationError((e.args or (_('Invalid credential.'),))[0])
|
|
except ValueError:
|
|
pass
|
|
return cred
|
|
|
|
def resolve_execution_environment(self):
|
|
"""
|
|
Project updates, themselves, will use the control plane execution environment.
|
|
Jobs using the project can use the default_environment, but the project updates
|
|
are not flexible enough to allow customizing the image they use.
|
|
"""
|
|
return get_control_plane_execution_environment()
|
|
|
|
def get_project_path(self, check_if_exists=True):
|
|
local_path = os.path.basename(self.local_path)
|
|
if local_path and not local_path.startswith('.'):
|
|
proj_path = os.path.join(settings.PROJECTS_ROOT, local_path)
|
|
if not check_if_exists or os.path.exists(smart_str(proj_path)):
|
|
return proj_path
|
|
|
|
def get_cache_path(self):
|
|
local_path = os.path.basename(self.local_path)
|
|
if local_path:
|
|
return os.path.join(settings.PROJECTS_ROOT, '.__awx_cache', local_path)
|
|
|
|
@property
|
|
def playbooks(self):
|
|
results = []
|
|
project_path = self.get_project_path()
|
|
if project_path:
|
|
for dirpath, dirnames, filenames in os.walk(smart_str(project_path), followlinks=settings.AWX_SHOW_PLAYBOOK_LINKS):
|
|
if skip_directory(dirpath):
|
|
continue
|
|
for filename in filenames:
|
|
playbook = could_be_playbook(project_path, dirpath, filename)
|
|
if playbook is not None:
|
|
results.append(smart_str(playbook))
|
|
return sorted(results, key=lambda x: smart_str(x).lower())
|
|
|
|
@property
|
|
def inventories(self):
|
|
results = []
|
|
project_path = self.get_project_path()
|
|
if project_path:
|
|
# Cap the number of results, because it could include lots
|
|
max_inventory_listing = 50
|
|
for dirpath, dirnames, filenames in os.walk(smart_str(project_path)):
|
|
if skip_directory(dirpath):
|
|
continue
|
|
for filename in filenames:
|
|
inv_path = could_be_inventory(project_path, dirpath, filename)
|
|
if inv_path is not None:
|
|
results.append(smart_str(inv_path))
|
|
if len(results) > max_inventory_listing:
|
|
break
|
|
if len(results) > max_inventory_listing:
|
|
break
|
|
return sorted(results, key=lambda x: smart_str(x).lower())
|
|
|
|
def get_lock_file(self):
|
|
"""
|
|
We want the project path in name only, we don't care if it exists or
|
|
not. This method will just append .lock onto the full directory path.
|
|
"""
|
|
proj_path = self.get_project_path(check_if_exists=False)
|
|
if not proj_path:
|
|
return None
|
|
return proj_path + '.lock'
|
|
|
|
|
|
class Project(UnifiedJobTemplate, ProjectOptions, ResourceMixin, CustomVirtualEnvMixin, RelatedJobsMixin):
|
|
"""
|
|
A project represents a playbook git repo that can access a set of inventories
|
|
"""
|
|
|
|
SOFT_UNIQUE_TOGETHER = [('polymorphic_ctype', 'name', 'organization')]
|
|
FIELDS_TO_PRESERVE_AT_COPY = ['labels', 'instance_groups', 'credentials']
|
|
FIELDS_TO_DISCARD_AT_COPY = ['local_path']
|
|
FIELDS_TRIGGER_UPDATE = frozenset(['scm_url', 'scm_branch', 'scm_type', 'scm_refspec'])
|
|
|
|
class Meta:
|
|
app_label = 'main'
|
|
ordering = ('id',)
|
|
|
|
default_environment = models.ForeignKey(
|
|
'ExecutionEnvironment',
|
|
null=True,
|
|
blank=True,
|
|
default=None,
|
|
on_delete=polymorphic.SET_NULL,
|
|
related_name='+',
|
|
help_text=_('The default execution environment for jobs run using this project.'),
|
|
)
|
|
scm_update_on_launch = models.BooleanField(
|
|
default=False,
|
|
help_text=_('Update the project when a job is launched that uses the project.'),
|
|
)
|
|
scm_update_cache_timeout = models.PositiveIntegerField(
|
|
default=0,
|
|
blank=True,
|
|
help_text=_('The number of seconds after the last project update ran that a new ' 'project update will be launched as a job dependency.'),
|
|
)
|
|
allow_override = models.BooleanField(
|
|
default=False,
|
|
help_text=_('Allow changing the SCM branch or revision in a job template ' 'that uses this project.'),
|
|
)
|
|
|
|
# credential (keys) used to validate content signature
|
|
signature_validation_credential = models.ForeignKey(
|
|
'Credential',
|
|
related_name='%(class)ss_signature_validation',
|
|
blank=True,
|
|
null=True,
|
|
default=None,
|
|
on_delete=models.SET_NULL,
|
|
help_text=_('An optional credential used for validating files in the project against unexpected changes.'),
|
|
)
|
|
|
|
scm_revision = models.CharField(
|
|
max_length=1024,
|
|
blank=True,
|
|
default='',
|
|
editable=False,
|
|
verbose_name=_('SCM Revision'),
|
|
help_text=_('The last revision fetched by a project update'),
|
|
)
|
|
|
|
playbook_files = JSONBlob(
|
|
default=list,
|
|
blank=True,
|
|
editable=False,
|
|
verbose_name=_('Playbook Files'),
|
|
help_text=_('List of playbooks found in the project'),
|
|
)
|
|
|
|
inventory_files = JSONBlob(
|
|
default=list,
|
|
blank=True,
|
|
editable=False,
|
|
verbose_name=_('Inventory Files'),
|
|
help_text=_('Suggested list of content that could be Ansible inventory in the project'),
|
|
)
|
|
|
|
admin_role = ImplicitRoleField(
|
|
parent_role=[
|
|
'organization.project_admin_role',
|
|
'singleton:' + ROLE_SINGLETON_SYSTEM_ADMINISTRATOR,
|
|
]
|
|
)
|
|
|
|
use_role = ImplicitRoleField(
|
|
parent_role='admin_role',
|
|
)
|
|
|
|
update_role = ImplicitRoleField(
|
|
parent_role='admin_role',
|
|
)
|
|
|
|
read_role = ImplicitRoleField(
|
|
parent_role=[
|
|
'organization.auditor_role',
|
|
'singleton:' + ROLE_SINGLETON_SYSTEM_AUDITOR,
|
|
'use_role',
|
|
'update_role',
|
|
]
|
|
)
|
|
|
|
@classmethod
|
|
def _get_unified_job_class(cls):
|
|
return ProjectUpdate
|
|
|
|
@classmethod
|
|
def _get_unified_job_field_names(cls):
|
|
return set(f.name for f in ProjectOptions._meta.fields) | set(['name', 'description', 'organization'])
|
|
|
|
def clean_organization(self):
|
|
if self.pk:
|
|
old_org_id = getattr(self, '_prior_values_store', {}).get('organization_id', None)
|
|
if self.organization_id != old_org_id and self.jobtemplates.exists():
|
|
raise ValidationError({'organization': _('Organization cannot be changed when in use by job templates.')})
|
|
return self.organization
|
|
|
|
def save(self, *args, **kwargs):
|
|
new_instance = not bool(self.pk)
|
|
pre_save_vals = getattr(self, '_prior_values_store', {})
|
|
# If update_fields has been specified, add our field names to it,
|
|
# if it hasn't been specified, then we're just doing a normal save.
|
|
update_fields = kwargs.get('update_fields', [])
|
|
self._skip_update = bool(kwargs.pop('skip_update', False))
|
|
# Create auto-generated local path if project uses SCM.
|
|
if self.pk and self.scm_type and not self.local_path.startswith('_'):
|
|
slug_name = slugify(str(self.name)).replace(u'-', u'_')
|
|
self.local_path = u'_%d__%s' % (int(self.pk), slug_name)
|
|
if 'local_path' not in update_fields:
|
|
update_fields.append('local_path')
|
|
# Do the actual save.
|
|
super(Project, self).save(*args, **kwargs)
|
|
if new_instance:
|
|
update_fields = []
|
|
# Generate local_path for SCM after initial save (so we have a PK).
|
|
if self.scm_type and not self.local_path.startswith('_'):
|
|
update_fields.append('local_path')
|
|
if update_fields:
|
|
from awx.main.signals import disable_activity_stream
|
|
|
|
with disable_activity_stream():
|
|
self.save(update_fields=update_fields, skip_update=self._skip_update)
|
|
# If we just created a new project with SCM, start the initial update.
|
|
# also update if certain fields have changed
|
|
relevant_change = any(pre_save_vals.get(fd_name, None) != self._prior_values_store.get(fd_name, None) for fd_name in self.FIELDS_TRIGGER_UPDATE)
|
|
if (relevant_change or new_instance) and (not self._skip_update) and self.scm_type:
|
|
self.update()
|
|
|
|
def _get_current_status(self):
|
|
if getattr(self, '_skip_update', False):
|
|
return self.status
|
|
if self.scm_type:
|
|
if self.current_job and self.current_job.status:
|
|
return self.current_job.status
|
|
elif not self.last_job:
|
|
return 'never updated'
|
|
# inherit the child job status on failure
|
|
elif self.last_job_failed:
|
|
return self.last_job.status
|
|
# Return the successful status
|
|
else:
|
|
return self.last_job.status
|
|
elif not self.get_project_path():
|
|
return 'missing'
|
|
else:
|
|
return 'ok'
|
|
|
|
def _get_last_job_run(self):
|
|
if self.scm_type and self.last_job:
|
|
return self.last_job.finished
|
|
else:
|
|
project_path = self.get_project_path()
|
|
if project_path:
|
|
try:
|
|
mtime = os.path.getmtime(smart_str(project_path))
|
|
dt = datetime.datetime.fromtimestamp(mtime)
|
|
return make_aware(dt, get_default_timezone())
|
|
except os.error:
|
|
pass
|
|
|
|
def _can_update(self):
|
|
return bool(self.scm_type)
|
|
|
|
def create_project_update(self, **kwargs):
|
|
return self.create_unified_job(**kwargs)
|
|
|
|
@property
|
|
def cache_timeout_blocked(self):
|
|
if not self.last_job_run:
|
|
return False
|
|
if (self.last_job_run + datetime.timedelta(seconds=self.scm_update_cache_timeout)) > now():
|
|
return True
|
|
return False
|
|
|
|
@property
|
|
def needs_update_on_launch(self):
|
|
if self.scm_type and self.scm_update_on_launch:
|
|
if not self.last_job_run:
|
|
return True
|
|
if (self.last_job_run + datetime.timedelta(seconds=self.scm_update_cache_timeout)) <= now():
|
|
return True
|
|
return False
|
|
|
|
@property
|
|
def cache_id(self):
|
|
return str(self.last_job_id)
|
|
|
|
@property
|
|
def notification_templates(self):
|
|
base_notification_templates = NotificationTemplate.objects
|
|
error_notification_templates = list(base_notification_templates.filter(unifiedjobtemplate_notification_templates_for_errors=self))
|
|
started_notification_templates = list(base_notification_templates.filter(unifiedjobtemplate_notification_templates_for_started=self))
|
|
success_notification_templates = list(base_notification_templates.filter(unifiedjobtemplate_notification_templates_for_success=self))
|
|
# Get Organization NotificationTemplates
|
|
if self.organization is not None:
|
|
error_notification_templates = set(
|
|
error_notification_templates + list(base_notification_templates.filter(organization_notification_templates_for_errors=self.organization))
|
|
)
|
|
started_notification_templates = set(
|
|
started_notification_templates + list(base_notification_templates.filter(organization_notification_templates_for_started=self.organization))
|
|
)
|
|
success_notification_templates = set(
|
|
success_notification_templates + list(base_notification_templates.filter(organization_notification_templates_for_success=self.organization))
|
|
)
|
|
return dict(error=list(error_notification_templates), started=list(started_notification_templates), success=list(success_notification_templates))
|
|
|
|
def get_absolute_url(self, request=None):
|
|
return reverse('api:project_detail', kwargs={'pk': self.pk}, request=request)
|
|
|
|
def get_reason_if_failed(self):
|
|
"""
|
|
If the project is in a failed or errored state, return a human-readable
|
|
error message explaining why. Otherwise return None.
|
|
|
|
This is used during validation in the serializer and also by
|
|
RunProjectUpdate/RunInventoryUpdate.
|
|
"""
|
|
|
|
if self.status not in ('error', 'failed'):
|
|
return None
|
|
|
|
latest_update = self.project_updates.last()
|
|
if latest_update is not None and latest_update.failed:
|
|
failed_validation_tasks = latest_update.project_update_events.filter(
|
|
event='runner_on_failed',
|
|
play="Perform project signature/checksum verification",
|
|
)
|
|
if failed_validation_tasks:
|
|
return _("Last project update failed due to signature validation failure.")
|
|
|
|
return _("Missing a revision to run due to failed project update.")
|
|
|
|
'''
|
|
RelatedJobsMixin
|
|
'''
|
|
|
|
def _get_related_jobs(self):
|
|
return UnifiedJob.objects.non_polymorphic().filter(models.Q(job__project=self) | models.Q(projectupdate__project=self))
|
|
|
|
def delete(self, *args, **kwargs):
|
|
paths_to_delete = (self.get_project_path(check_if_exists=False), self.get_cache_path())
|
|
r = super(Project, self).delete(*args, **kwargs)
|
|
for path_to_delete in paths_to_delete:
|
|
if self.scm_type and path_to_delete: # non-manual, concrete path
|
|
from awx.main.tasks.system import delete_project_files
|
|
|
|
delete_project_files.delay(path_to_delete)
|
|
return r
|
|
|
|
|
|
class ProjectUpdate(UnifiedJob, ProjectOptions, JobNotificationMixin, TaskManagerProjectUpdateMixin):
|
|
"""
|
|
Internal job for tracking project updates from SCM.
|
|
"""
|
|
|
|
class Meta:
|
|
app_label = 'main'
|
|
|
|
project = models.ForeignKey(
|
|
'Project',
|
|
related_name='project_updates',
|
|
on_delete=models.CASCADE,
|
|
editable=False,
|
|
)
|
|
|
|
job_type = models.CharField(
|
|
max_length=64,
|
|
choices=PROJECT_UPDATE_JOB_TYPE_CHOICES,
|
|
default='check',
|
|
)
|
|
job_tags = models.CharField(
|
|
max_length=1024,
|
|
blank=True,
|
|
default='',
|
|
help_text=_('Parts of the project update playbook that will be run.'),
|
|
)
|
|
scm_revision = models.CharField(
|
|
max_length=1024,
|
|
blank=True,
|
|
default='',
|
|
editable=False,
|
|
verbose_name=_('SCM Revision'),
|
|
help_text=_('The SCM Revision discovered by this update for the given project and branch.'),
|
|
)
|
|
|
|
def _set_default_dependencies_processed(self):
|
|
self.dependencies_processed = True
|
|
|
|
def _get_parent_field_name(self):
|
|
return 'project'
|
|
|
|
def _update_parent_instance(self):
|
|
if not self.project:
|
|
return # no parent instance to update
|
|
if self.job_type == PERM_INVENTORY_DEPLOY:
|
|
# Do not update project status if this is sync job
|
|
# unless no other updates have happened or started
|
|
first_update = False
|
|
if self.project.status == 'never updated' and self.status == 'running':
|
|
first_update = True
|
|
elif self.project.current_job == self:
|
|
first_update = True
|
|
if not first_update:
|
|
return
|
|
return super(ProjectUpdate, self)._update_parent_instance()
|
|
|
|
@classmethod
|
|
def _get_task_class(cls):
|
|
from awx.main.tasks.jobs import RunProjectUpdate
|
|
|
|
return RunProjectUpdate
|
|
|
|
def _global_timeout_setting(self):
|
|
return 'DEFAULT_PROJECT_UPDATE_TIMEOUT'
|
|
|
|
def is_blocked_by(self, obj):
|
|
if type(obj) == ProjectUpdate:
|
|
if self.project == obj.project:
|
|
return True
|
|
if type(obj) == Job:
|
|
if self.project == obj.project:
|
|
return True
|
|
return False
|
|
|
|
def websocket_emit_data(self):
|
|
websocket_data = super(ProjectUpdate, self).websocket_emit_data()
|
|
websocket_data.update(dict(project_id=self.project.id))
|
|
return websocket_data
|
|
|
|
@property
|
|
def event_class(self):
|
|
if self.has_unpartitioned_events:
|
|
return UnpartitionedProjectUpdateEvent
|
|
return ProjectUpdateEvent
|
|
|
|
def _get_task_impact(self):
|
|
return 0 if self.job_type == 'run' else 1
|
|
|
|
@property
|
|
def result_stdout(self):
|
|
return self._result_stdout_raw(redact_sensitive=True, escape_ascii=True)
|
|
|
|
@property
|
|
def result_stdout_raw(self):
|
|
return self._result_stdout_raw(redact_sensitive=True)
|
|
|
|
@property
|
|
def branch_override(self):
|
|
"""Whether a branch other than the project default is used."""
|
|
if not self.project:
|
|
return True
|
|
return bool(self.scm_branch and self.scm_branch != self.project.scm_branch)
|
|
|
|
@property
|
|
def cache_id(self):
|
|
if self.branch_override or self.job_type == 'check' or (not self.project):
|
|
return str(self.id)
|
|
return self.project.cache_id
|
|
|
|
def result_stdout_raw_limited(self, start_line=0, end_line=None, redact_sensitive=True):
|
|
return self._result_stdout_raw_limited(start_line, end_line, redact_sensitive=redact_sensitive)
|
|
|
|
def result_stdout_limited(self, start_line=0, end_line=None, redact_sensitive=True):
|
|
return self._result_stdout_raw_limited(start_line, end_line, redact_sensitive=redact_sensitive, escape_ascii=True)
|
|
|
|
def get_absolute_url(self, request=None):
|
|
return reverse('api:project_update_detail', kwargs={'pk': self.pk}, request=request)
|
|
|
|
def get_ui_url(self):
|
|
return urlparse.urljoin(settings.TOWER_URL_BASE, "/#/jobs/project/{}".format(self.pk))
|
|
|
|
def cancel(self, job_explanation=None, is_chain=False):
|
|
res = super(ProjectUpdate, self).cancel(job_explanation=job_explanation, is_chain=is_chain)
|
|
if res and self.launch_type != 'sync':
|
|
for inv_src in self.scm_inventory_updates.filter(status='running'):
|
|
inv_src.cancel(job_explanation='Source project update `{}` was canceled.'.format(self.name))
|
|
return res
|
|
|
|
'''
|
|
JobNotificationMixin
|
|
'''
|
|
|
|
def get_notification_templates(self):
|
|
return self.project.notification_templates
|
|
|
|
def get_notification_friendly_name(self):
|
|
return "Project Update"
|
|
|
|
def save(self, *args, **kwargs):
|
|
added_update_fields = []
|
|
if not self.job_tags:
|
|
job_tags = ['update_{}'.format(self.scm_type), 'install_roles', 'install_collections']
|
|
if self.project.signature_validation_credential is not None:
|
|
credential_type = self.project.signature_validation_credential.credential_type.namespace
|
|
job_tags.append(f'validation_{credential_type}')
|
|
job_tags.append('validation_checksum_manifest')
|
|
self.job_tags = ','.join(job_tags)
|
|
added_update_fields.append('job_tags')
|
|
if self.scm_delete_on_update and 'delete' not in self.job_tags and self.job_type == 'check':
|
|
self.job_tags = ','.join([self.job_tags, 'delete'])
|
|
added_update_fields.append('job_tags')
|
|
elif (not self.scm_delete_on_update) and 'delete' in self.job_tags:
|
|
job_tags = self.job_tags.split(',')
|
|
job_tags.remove('delete')
|
|
self.job_tags = ','.join(job_tags)
|
|
added_update_fields.append('job_tags')
|
|
if 'update_fields' in kwargs:
|
|
kwargs['update_fields'].extend(added_update_fields)
|
|
return super(ProjectUpdate, self).save(*args, **kwargs)
|