From 91bfed3d504db0291c36862fe5ba3619f381a1a1 Mon Sep 17 00:00:00 2001 From: chris meyers Date: Wed, 21 Mar 2018 13:41:48 -0400 Subject: [PATCH] handle instance group names unicode --- awx/main/scheduler/task_manager.py | 29 +++++++++++++++-------------- 1 file changed, 15 insertions(+), 14 deletions(-) diff --git a/awx/main/scheduler/task_manager.py b/awx/main/scheduler/task_manager.py index ea46146b60..861d098a02 100644 --- a/awx/main/scheduler/task_manager.py +++ b/awx/main/scheduler/task_manager.py @@ -6,6 +6,7 @@ from datetime import datetime, timedelta import logging import uuid import json +import six from sets import Set # Django @@ -422,50 +423,50 @@ class TaskManager(): def process_dependencies(self, dependent_task, dependency_tasks): for task in dependency_tasks: if self.is_job_blocked(task): - logger.debug("Dependent %s is blocked from running", task.log_format) + logger.debug(six.text_type("Dependent {} is blocked from running").format(task.log_format)) continue preferred_instance_groups = task.preferred_instance_groups found_acceptable_queue = False for rampart_group in preferred_instance_groups: if self.get_remaining_capacity(rampart_group.name) <= 0: - logger.debug("Skipping group %s capacity <= 0", rampart_group.name) + logger.debug(six.text_type("Skipping group {} capacity <= 0").format(rampart_group.name)) continue if not self.would_exceed_capacity(task, rampart_group.name): - logger.debug("Starting dependent %s in group %s", task.log_format, rampart_group.name) + logger.debug(six.text_type("Starting dependent {} in group {}").format(task.log_format, rampart_group.name)) self.graph[rampart_group.name]['graph'].add_job(task) tasks_to_fail = filter(lambda t: t != task, dependency_tasks) tasks_to_fail += [dependent_task] self.start_task(task, rampart_group, tasks_to_fail) found_acceptable_queue = True if not found_acceptable_queue: - logger.debug("Dependent %s couldn't be scheduled on graph, waiting for next cycle", task.log_format) + logger.debug(six.text_type("Dependent {} couldn't be scheduled on graph, waiting for next cycle").format(task.log_format)) def process_pending_tasks(self, pending_tasks): for task in pending_tasks: self.process_dependencies(task, self.generate_dependencies(task)) if self.is_job_blocked(task): - logger.debug("%s is blocked from running", task.log_format) + logger.debug(six.text_type("{} is blocked from running").format(task.log_format)) continue preferred_instance_groups = task.preferred_instance_groups found_acceptable_queue = False for rampart_group in preferred_instance_groups: remaining_capacity = self.get_remaining_capacity(rampart_group.name) if remaining_capacity <= 0: - logger.debug("Skipping group %s, remaining_capacity %s <= 0", - rampart_group.name, remaining_capacity) + logger.debug(six.text_type("Skipping group {}, remaining_capacity {} <= 0").format( + rampart_group.name, remaining_capacity)) continue if not self.would_exceed_capacity(task, rampart_group.name): - logger.debug("Starting %s in group %s (remaining_capacity=%s)", - task.log_format, rampart_group.name, remaining_capacity) + logger.debug(six.text_type("Starting {} in group {} (remaining_capacity={})").format( + task.log_format, rampart_group.name, remaining_capacity)) self.graph[rampart_group.name]['graph'].add_job(task) self.start_task(task, rampart_group, task.get_jobs_fail_chain()) found_acceptable_queue = True break else: - logger.debug("Not enough capacity to run %s on %s (remaining_capacity=%s)", - task.log_format, rampart_group.name, remaining_capacity) + logger.debug(six.text_type("Not enough capacity to run {} on {} (remaining_capacity={})").format( + task.log_format, rampart_group.name, remaining_capacity)) if not found_acceptable_queue: - logger.debug("%s couldn't be scheduled on graph, waiting for next cycle", task.log_format) + logger.debug(six.text_type("{} couldn't be scheduled on graph, waiting for next cycle").format(task.log_format)) def fail_jobs_if_not_in_celery(self, node_jobs, active_tasks, celery_task_start_time, isolated=False): @@ -576,9 +577,9 @@ class TaskManager(): return (task.task_impact + current_capacity > capacity_total) def consume_capacity(self, task, instance_group): - logger.debug('%s consumed %s capacity units from %s with prior total of %s', + logger.debug(six.text_type('{} consumed {} capacity units from {} with prior total of {}').format( task.log_format, task.task_impact, instance_group, - self.graph[instance_group]['consumed_capacity']) + self.graph[instance_group]['consumed_capacity'])) self.graph[instance_group]['consumed_capacity'] += task.task_impact def get_remaining_capacity(self, instance_group):