Wait 60 seconds before scaling down a worker

This commit is contained in:
Alan Rominger 2022-07-27 16:54:03 -04:00
parent 85a5b58d18
commit a9170236e1
No known key found for this signature in database
GPG Key ID: C2D7EAAA12B63559
2 changed files with 20 additions and 3 deletions

View File

@ -71,9 +71,11 @@ class PoolWorker(object):
self.messages_finished = 0
self.managed_tasks = collections.OrderedDict()
self.finished = MPQueue(queue_size) if self.track_managed_tasks else NoOpResultQueue()
self.last_finished = None
self.queue = MPQueue(queue_size)
self.process = Process(target=target, args=(self.queue, self.finished) + args)
self.process.daemon = True
self.scale_down_in = settings.DISPATCHER_SCALE_DOWN_WAIT_TIME
def start(self):
self.process.start()
@ -144,6 +146,9 @@ class PoolWorker(object):
# state of which events are *currently* being processed.
logger.warning('Event UUID {} appears to be have been duplicated.'.format(uuid))
if finished:
self.last_finished = time.time()
@property
def current_task(self):
if not self.track_managed_tasks:
@ -189,6 +194,14 @@ class PoolWorker(object):
def idle(self):
return not self.busy
@property
def lazy(self):
if self.busy:
return False
if self.last_finished is None:
return True
return time.time() - self.last_finished > self.scale_down_in
class StatefulPoolWorker(PoolWorker):
@ -249,7 +262,7 @@ class WorkerPool(object):
except Exception:
logger.exception('could not fork')
else:
logger.debug('scaling up worker pid:{}'.format(worker.pid))
logger.info(f'scaling up worker pid:{worker.pid} total:{len(self.workers)}')
return idx, worker
def debug(self, *args, **kwargs):
@ -385,12 +398,12 @@ class AutoscalePool(WorkerPool):
logger.exception('failed to reap job UUID {}'.format(w.current_task['uuid']))
orphaned.extend(w.orphaned_tasks)
self.workers.remove(w)
elif w.idle and len(self.workers) > self.min_workers:
elif w.lazy and len(self.workers) > self.min_workers:
# the process has an empty queue (it's idle) and we have
# more processes in the pool than we need (> min)
# send this process a message so it will exit gracefully
# at the next opportunity
logger.debug('scaling down worker pid:{}'.format(w.pid))
logger.info(f'scaling down worker pid:{w.pid} from:{len(self.workers)}')
w.quit()
self.workers.remove(w)
if w.alive:

View File

@ -438,6 +438,10 @@ EXECUTION_NODE_REMEDIATION_CHECKS = 60 * 30 # once every 30 minutes check if an
# Amount of time dispatcher will try to reconnect to database for jobs and consuming new work
DISPATCHER_DB_DOWNTOWN_TOLLERANCE = 40
# Minimum time to wait after last job finished before scaling down a worker
# A higher value will free up memory more agressively, but a lower value will require less forking
DISPATCHER_SCALE_DOWN_WAIT_TIME = 60
BROKER_URL = 'unix:///var/run/redis/redis.sock'
CELERYBEAT_SCHEDULE = {
'tower_scheduler': {'task': 'awx.main.tasks.system.awx_periodic_scheduler', 'schedule': timedelta(seconds=30), 'options': {'expires': 20}},