mirror of
https://github.com/ansible/awx.git
synced 2026-04-14 14:39:26 -02:30
Compare commits
127 Commits
21.9.0
...
13089-Sche
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
26a947ed31 | ||
|
|
b99a434dee | ||
|
|
6cee99a9f9 | ||
|
|
ee509aea56 | ||
|
|
b5452a48f8 | ||
|
|
0c980fa7d5 | ||
|
|
e34ce8c795 | ||
|
|
3543644e0e | ||
|
|
36c0d07b30 | ||
|
|
239827a9cf | ||
|
|
ac9871b36f | ||
|
|
f739908ccf | ||
|
|
cf1ec07eab | ||
|
|
d968b648de | ||
|
|
5dd0eab806 | ||
|
|
41f3f381ec | ||
|
|
ac8cff75ce | ||
|
|
94b34b801c | ||
|
|
8f6849fc22 | ||
|
|
821b1701bf | ||
|
|
b7f2825909 | ||
|
|
e87e041a2a | ||
|
|
cc336e791c | ||
|
|
c2a3c3b285 | ||
|
|
7b8dcc98e7 | ||
|
|
d5011492bf | ||
|
|
e363ddf470 | ||
|
|
987709cdb3 | ||
|
|
f04ac3c798 | ||
|
|
71a6baccdb | ||
|
|
d07076b686 | ||
|
|
7129f3e8cd | ||
|
|
df61a5cea1 | ||
|
|
a4b950f79b | ||
|
|
8be739d255 | ||
|
|
ca54195099 | ||
|
|
f0fcfdde39 | ||
|
|
80b1ba4a35 | ||
|
|
51f8e362dc | ||
|
|
737d6d8c8b | ||
|
|
beaf6b6058 | ||
|
|
aad1fbcef8 | ||
|
|
0b96d617ac | ||
|
|
fe768a159b | ||
|
|
c1ebea858b | ||
|
|
da9b8135e8 | ||
|
|
76cecf3f6b | ||
|
|
7b2938f515 | ||
|
|
916b5642d2 | ||
|
|
e524d3df3e | ||
|
|
01e9a611ea | ||
|
|
ef29589940 | ||
|
|
cec2d2dfb9 | ||
|
|
15b7ad3570 | ||
|
|
36ff9cbc6d | ||
|
|
ed74d80ecb | ||
|
|
a0b8215c06 | ||
|
|
f88b993b18 | ||
|
|
4a7f4d0ed4 | ||
|
|
6e08c3567f | ||
|
|
adbcb5c5e4 | ||
|
|
8054c6aedc | ||
|
|
58734a33c4 | ||
|
|
2832f28014 | ||
|
|
e5057691ee | ||
|
|
a0cfd8501c | ||
|
|
99b643bd77 | ||
|
|
305b39d8e5 | ||
|
|
bb047baeba | ||
|
|
9637aad37e | ||
|
|
fbc06ec623 | ||
|
|
57430afc55 | ||
|
|
7aae7e8ed4 | ||
|
|
a67d107a58 | ||
|
|
642003e207 | ||
|
|
ec7e2284df | ||
|
|
ff7facdfa2 | ||
|
|
6df4e62132 | ||
|
|
6289bfb639 | ||
|
|
95e4b2064f | ||
|
|
48eba60be4 | ||
|
|
c7efa8b4e0 | ||
|
|
657b5cb1aa | ||
|
|
06daebbecf | ||
|
|
fb37f22bf4 | ||
|
|
71f326b705 | ||
|
|
6508ab4a33 | ||
|
|
bf871bd427 | ||
|
|
e403c603d6 | ||
|
|
4b7b3c7c7d | ||
|
|
1cdd2cad67 | ||
|
|
86856f242a | ||
|
|
65c3db8cb8 | ||
|
|
7fa9dcbc2a | ||
|
|
7cfb957de3 | ||
|
|
d0d467e863 | ||
|
|
eaccf32aa3 | ||
|
|
a8fdb22ab3 | ||
|
|
ae79f94a48 | ||
|
|
40499a4084 | ||
|
|
b36fa93005 | ||
|
|
8839b4e90b | ||
|
|
7866135d6c | ||
|
|
fe48dc412f | ||
|
|
3a25c4221f | ||
|
|
7e1be3ef94 | ||
|
|
b2f8ca09ba | ||
|
|
c7692f5c56 | ||
|
|
3b24afa7f2 | ||
|
|
2b3f3e2043 | ||
|
|
68614b83c0 | ||
|
|
a1edc75c11 | ||
|
|
4b0e7a5cde | ||
|
|
f0481d0a60 | ||
|
|
d34f6af830 | ||
|
|
8b9db837ca | ||
|
|
f9bb26ad33 | ||
|
|
271613b86d | ||
|
|
ac57f5cb28 | ||
|
|
c39172f516 | ||
|
|
9b047c2af6 | ||
|
|
f0d6bc0dc8 | ||
|
|
7590301ae7 | ||
|
|
878035c13b | ||
|
|
2cc971a43f | ||
|
|
9d77c54612 | ||
|
|
ef651a3a21 |
10
.github/triage_replies.md
vendored
10
.github/triage_replies.md
vendored
@@ -53,6 +53,16 @@ https://github.com/ansible/awx/#get-involved \
|
|||||||
Thank you once again for this and your interest in AWX!
|
Thank you once again for this and your interest in AWX!
|
||||||
|
|
||||||
|
|
||||||
|
### Red Hat Support Team
|
||||||
|
- Hi! \
|
||||||
|
\
|
||||||
|
It appears that you are using an RPM build for RHEL. Please reach out to the Red Hat support team and submit a ticket. \
|
||||||
|
\
|
||||||
|
Here is the link to do so: \
|
||||||
|
\
|
||||||
|
https://access.redhat.com/support \
|
||||||
|
\
|
||||||
|
Thank you for your submission and for supporting AWX!
|
||||||
|
|
||||||
|
|
||||||
## Common
|
## Common
|
||||||
|
|||||||
20
.github/workflows/ci.yml
vendored
20
.github/workflows/ci.yml
vendored
@@ -2,6 +2,7 @@
|
|||||||
name: CI
|
name: CI
|
||||||
env:
|
env:
|
||||||
BRANCH: ${{ github.base_ref || 'devel' }}
|
BRANCH: ${{ github.base_ref || 'devel' }}
|
||||||
|
LC_ALL: "C.UTF-8" # prevent ERROR: Ansible could not initialize the preferred locale: unsupported locale setting
|
||||||
on:
|
on:
|
||||||
pull_request:
|
pull_request:
|
||||||
jobs:
|
jobs:
|
||||||
@@ -144,3 +145,22 @@ jobs:
|
|||||||
env:
|
env:
|
||||||
AWX_TEST_IMAGE: awx
|
AWX_TEST_IMAGE: awx
|
||||||
AWX_TEST_VERSION: ci
|
AWX_TEST_VERSION: ci
|
||||||
|
|
||||||
|
collection-sanity:
|
||||||
|
name: awx_collection sanity
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
strategy:
|
||||||
|
fail-fast: false
|
||||||
|
steps:
|
||||||
|
- uses: actions/checkout@v2
|
||||||
|
|
||||||
|
# The containers that GitHub Actions use have Ansible installed, so upgrade to make sure we have the latest version.
|
||||||
|
- name: Upgrade ansible-core
|
||||||
|
run: python3 -m pip install --upgrade ansible-core
|
||||||
|
|
||||||
|
- name: Run sanity tests
|
||||||
|
run: make test_collection_sanity
|
||||||
|
env:
|
||||||
|
# needed due to cgroupsv2. This is fixed, but a stable release
|
||||||
|
# with the fix has not been made yet.
|
||||||
|
ANSIBLE_TEST_PREFER_PODMAN: 1
|
||||||
|
|||||||
2
.github/workflows/devel_images.yml
vendored
2
.github/workflows/devel_images.yml
vendored
@@ -1,5 +1,7 @@
|
|||||||
---
|
---
|
||||||
name: Build/Push Development Images
|
name: Build/Push Development Images
|
||||||
|
env:
|
||||||
|
LC_ALL: "C.UTF-8" # prevent ERROR: Ansible could not initialize the preferred locale: unsupported locale setting
|
||||||
on:
|
on:
|
||||||
push:
|
push:
|
||||||
branches:
|
branches:
|
||||||
|
|||||||
5
.github/workflows/e2e_test.yml
vendored
5
.github/workflows/e2e_test.yml
vendored
@@ -1,5 +1,8 @@
|
|||||||
---
|
---
|
||||||
name: E2E Tests
|
name: E2E Tests
|
||||||
|
env:
|
||||||
|
LC_ALL: "C.UTF-8" # prevent ERROR: Ansible could not initialize the preferred locale: unsupported locale setting
|
||||||
|
|
||||||
on:
|
on:
|
||||||
pull_request_target:
|
pull_request_target:
|
||||||
types: [labeled]
|
types: [labeled]
|
||||||
@@ -104,5 +107,3 @@ jobs:
|
|||||||
with:
|
with:
|
||||||
name: AWX-logs-${{ matrix.job }}
|
name: AWX-logs-${{ matrix.job }}
|
||||||
path: make-docker-compose-output.log
|
path: make-docker-compose-output.log
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
@@ -1,5 +1,7 @@
|
|||||||
---
|
---
|
||||||
name: Feature branch deletion cleanup
|
name: Feature branch deletion cleanup
|
||||||
|
env:
|
||||||
|
LC_ALL: "C.UTF-8" # prevent ERROR: Ansible could not initialize the preferred locale: unsupported locale setting
|
||||||
on:
|
on:
|
||||||
delete:
|
delete:
|
||||||
branches:
|
branches:
|
||||||
|
|||||||
4
.github/workflows/promote.yml
vendored
4
.github/workflows/promote.yml
vendored
@@ -1,5 +1,9 @@
|
|||||||
---
|
---
|
||||||
name: Promote Release
|
name: Promote Release
|
||||||
|
|
||||||
|
env:
|
||||||
|
LC_ALL: "C.UTF-8" # prevent ERROR: Ansible could not initialize the preferred locale: unsupported locale setting
|
||||||
|
|
||||||
on:
|
on:
|
||||||
release:
|
release:
|
||||||
types: [published]
|
types: [published]
|
||||||
|
|||||||
4
.github/workflows/stage.yml
vendored
4
.github/workflows/stage.yml
vendored
@@ -1,5 +1,9 @@
|
|||||||
---
|
---
|
||||||
name: Stage Release
|
name: Stage Release
|
||||||
|
|
||||||
|
env:
|
||||||
|
LC_ALL: "C.UTF-8" # prevent ERROR: Ansible could not initialize the preferred locale: unsupported locale setting
|
||||||
|
|
||||||
on:
|
on:
|
||||||
workflow_dispatch:
|
workflow_dispatch:
|
||||||
inputs:
|
inputs:
|
||||||
|
|||||||
4
.github/workflows/upload_schema.yml
vendored
4
.github/workflows/upload_schema.yml
vendored
@@ -1,5 +1,9 @@
|
|||||||
---
|
---
|
||||||
name: Upload API Schema
|
name: Upload API Schema
|
||||||
|
|
||||||
|
env:
|
||||||
|
LC_ALL: "C.UTF-8" # prevent ERROR: Ansible could not initialize the preferred locale: unsupported locale setting
|
||||||
|
|
||||||
on:
|
on:
|
||||||
push:
|
push:
|
||||||
branches:
|
branches:
|
||||||
|
|||||||
51
Makefile
51
Makefile
@@ -6,7 +6,20 @@ CHROMIUM_BIN=/tmp/chrome-linux/chrome
|
|||||||
GIT_BRANCH ?= $(shell git rev-parse --abbrev-ref HEAD)
|
GIT_BRANCH ?= $(shell git rev-parse --abbrev-ref HEAD)
|
||||||
MANAGEMENT_COMMAND ?= awx-manage
|
MANAGEMENT_COMMAND ?= awx-manage
|
||||||
VERSION := $(shell $(PYTHON) tools/scripts/scm_version.py)
|
VERSION := $(shell $(PYTHON) tools/scripts/scm_version.py)
|
||||||
COLLECTION_VERSION := $(shell $(PYTHON) tools/scripts/scm_version.py | cut -d . -f 1-3)
|
|
||||||
|
# ansible-test requires semver compatable version, so we allow overrides to hack it
|
||||||
|
COLLECTION_VERSION ?= $(shell $(PYTHON) tools/scripts/scm_version.py | cut -d . -f 1-3)
|
||||||
|
# args for the ansible-test sanity command
|
||||||
|
COLLECTION_SANITY_ARGS ?= --docker
|
||||||
|
# collection unit testing directories
|
||||||
|
COLLECTION_TEST_DIRS ?= awx_collection/test/awx
|
||||||
|
# collection integration test directories (defaults to all)
|
||||||
|
COLLECTION_TEST_TARGET ?=
|
||||||
|
# args for collection install
|
||||||
|
COLLECTION_PACKAGE ?= awx
|
||||||
|
COLLECTION_NAMESPACE ?= awx
|
||||||
|
COLLECTION_INSTALL = ~/.ansible/collections/ansible_collections/$(COLLECTION_NAMESPACE)/$(COLLECTION_PACKAGE)
|
||||||
|
COLLECTION_TEMPLATE_VERSION ?= false
|
||||||
|
|
||||||
# NOTE: This defaults the container image version to the branch that's active
|
# NOTE: This defaults the container image version to the branch that's active
|
||||||
COMPOSE_TAG ?= $(GIT_BRANCH)
|
COMPOSE_TAG ?= $(GIT_BRANCH)
|
||||||
@@ -34,7 +47,7 @@ RECEPTOR_IMAGE ?= quay.io/ansible/receptor:devel
|
|||||||
SRC_ONLY_PKGS ?= cffi,pycparser,psycopg2,twilio
|
SRC_ONLY_PKGS ?= cffi,pycparser,psycopg2,twilio
|
||||||
# These should be upgraded in the AWX and Ansible venv before attempting
|
# These should be upgraded in the AWX and Ansible venv before attempting
|
||||||
# to install the actual requirements
|
# to install the actual requirements
|
||||||
VENV_BOOTSTRAP ?= pip==21.2.4 setuptools==58.2.0 setuptools_scm[toml]==6.4.2 wheel==0.36.2
|
VENV_BOOTSTRAP ?= pip==21.2.4 setuptools==65.6.3 setuptools_scm[toml]==7.0.5 wheel==0.38.4
|
||||||
|
|
||||||
NAME ?= awx
|
NAME ?= awx
|
||||||
|
|
||||||
@@ -288,19 +301,13 @@ test:
|
|||||||
cd awxkit && $(VENV_BASE)/awx/bin/tox -re py3
|
cd awxkit && $(VENV_BASE)/awx/bin/tox -re py3
|
||||||
awx-manage check_migrations --dry-run --check -n 'missing_migration_file'
|
awx-manage check_migrations --dry-run --check -n 'missing_migration_file'
|
||||||
|
|
||||||
COLLECTION_TEST_DIRS ?= awx_collection/test/awx
|
|
||||||
COLLECTION_TEST_TARGET ?=
|
|
||||||
COLLECTION_PACKAGE ?= awx
|
|
||||||
COLLECTION_NAMESPACE ?= awx
|
|
||||||
COLLECTION_INSTALL = ~/.ansible/collections/ansible_collections/$(COLLECTION_NAMESPACE)/$(COLLECTION_PACKAGE)
|
|
||||||
COLLECTION_TEMPLATE_VERSION ?= false
|
|
||||||
|
|
||||||
test_collection:
|
test_collection:
|
||||||
rm -f $(shell ls -d $(VENV_BASE)/awx/lib/python* | head -n 1)/no-global-site-packages.txt
|
rm -f $(shell ls -d $(VENV_BASE)/awx/lib/python* | head -n 1)/no-global-site-packages.txt
|
||||||
if [ "$(VENV_BASE)" ]; then \
|
if [ "$(VENV_BASE)" ]; then \
|
||||||
. $(VENV_BASE)/awx/bin/activate; \
|
. $(VENV_BASE)/awx/bin/activate; \
|
||||||
fi && \
|
fi && \
|
||||||
pip install ansible-core && \
|
if ! [ -x "$(shell command -v ansible-playbook)" ]; then pip install ansible-core; fi
|
||||||
|
ansible --version
|
||||||
py.test $(COLLECTION_TEST_DIRS) -v
|
py.test $(COLLECTION_TEST_DIRS) -v
|
||||||
# The python path needs to be modified so that the tests can find Ansible within the container
|
# The python path needs to be modified so that the tests can find Ansible within the container
|
||||||
# First we will use anything expility set as PYTHONPATH
|
# First we will use anything expility set as PYTHONPATH
|
||||||
@@ -330,8 +337,13 @@ install_collection: build_collection
|
|||||||
rm -rf $(COLLECTION_INSTALL)
|
rm -rf $(COLLECTION_INSTALL)
|
||||||
ansible-galaxy collection install awx_collection_build/$(COLLECTION_NAMESPACE)-$(COLLECTION_PACKAGE)-$(COLLECTION_VERSION).tar.gz
|
ansible-galaxy collection install awx_collection_build/$(COLLECTION_NAMESPACE)-$(COLLECTION_PACKAGE)-$(COLLECTION_VERSION).tar.gz
|
||||||
|
|
||||||
test_collection_sanity: install_collection
|
test_collection_sanity:
|
||||||
cd $(COLLECTION_INSTALL) && ansible-test sanity
|
rm -rf awx_collection_build/
|
||||||
|
rm -rf $(COLLECTION_INSTALL)
|
||||||
|
if ! [ -x "$(shell command -v ansible-test)" ]; then pip install ansible-core; fi
|
||||||
|
ansible --version
|
||||||
|
COLLECTION_VERSION=1.0.0 make install_collection
|
||||||
|
cd $(COLLECTION_INSTALL) && ansible-test sanity $(COLLECTION_SANITY_ARGS)
|
||||||
|
|
||||||
test_collection_integration: install_collection
|
test_collection_integration: install_collection
|
||||||
cd $(COLLECTION_INSTALL) && ansible-test integration $(COLLECTION_TEST_TARGET)
|
cd $(COLLECTION_INSTALL) && ansible-test integration $(COLLECTION_TEST_TARGET)
|
||||||
@@ -389,18 +401,18 @@ $(UI_BUILD_FLAG_FILE):
|
|||||||
$(PYTHON) tools/scripts/compilemessages.py
|
$(PYTHON) tools/scripts/compilemessages.py
|
||||||
$(NPM_BIN) --prefix awx/ui --loglevel warn run compile-strings
|
$(NPM_BIN) --prefix awx/ui --loglevel warn run compile-strings
|
||||||
$(NPM_BIN) --prefix awx/ui --loglevel warn run build
|
$(NPM_BIN) --prefix awx/ui --loglevel warn run build
|
||||||
mkdir -p /var/lib/awx/public/static/css
|
|
||||||
mkdir -p /var/lib/awx/public/static/js
|
|
||||||
mkdir -p /var/lib/awx/public/static/media
|
|
||||||
cp -r awx/ui/build/static/css/* /var/lib/awx/public/static/css
|
|
||||||
cp -r awx/ui/build/static/js/* /var/lib/awx/public/static/js
|
|
||||||
cp -r awx/ui/build/static/media/* /var/lib/awx/public/static/media
|
|
||||||
touch $@
|
touch $@
|
||||||
|
|
||||||
ui-release: $(UI_BUILD_FLAG_FILE)
|
ui-release: $(UI_BUILD_FLAG_FILE)
|
||||||
|
|
||||||
ui-devel: awx/ui/node_modules
|
ui-devel: awx/ui/node_modules
|
||||||
@$(MAKE) -B $(UI_BUILD_FLAG_FILE)
|
@$(MAKE) -B $(UI_BUILD_FLAG_FILE)
|
||||||
|
mkdir -p /var/lib/awx/public/static/css
|
||||||
|
mkdir -p /var/lib/awx/public/static/js
|
||||||
|
mkdir -p /var/lib/awx/public/static/media
|
||||||
|
cp -r awx/ui/build/static/css/* /var/lib/awx/public/static/css
|
||||||
|
cp -r awx/ui/build/static/js/* /var/lib/awx/public/static/js
|
||||||
|
cp -r awx/ui/build/static/media/* /var/lib/awx/public/static/media
|
||||||
|
|
||||||
ui-devel-instrumented: awx/ui/node_modules
|
ui-devel-instrumented: awx/ui/node_modules
|
||||||
$(NPM_BIN) --prefix awx/ui --loglevel warn run start-instrumented
|
$(NPM_BIN) --prefix awx/ui --loglevel warn run start-instrumented
|
||||||
@@ -593,13 +605,12 @@ pot: $(UI_BUILD_FLAG_FILE)
|
|||||||
po: $(UI_BUILD_FLAG_FILE)
|
po: $(UI_BUILD_FLAG_FILE)
|
||||||
$(NPM_BIN) --prefix awx/ui --loglevel warn run extract-strings -- --clean
|
$(NPM_BIN) --prefix awx/ui --loglevel warn run extract-strings -- --clean
|
||||||
|
|
||||||
LANG = "en_us"
|
|
||||||
## generate API django .pot .po
|
## generate API django .pot .po
|
||||||
messages:
|
messages:
|
||||||
@if [ "$(VENV_BASE)" ]; then \
|
@if [ "$(VENV_BASE)" ]; then \
|
||||||
. $(VENV_BASE)/awx/bin/activate; \
|
. $(VENV_BASE)/awx/bin/activate; \
|
||||||
fi; \
|
fi; \
|
||||||
$(PYTHON) manage.py makemessages -l $(LANG) --keep-pot
|
$(PYTHON) manage.py makemessages -l en_us --keep-pot
|
||||||
|
|
||||||
print-%:
|
print-%:
|
||||||
@echo $($*)
|
@echo $($*)
|
||||||
|
|||||||
@@ -113,7 +113,7 @@ from awx.main.utils import (
|
|||||||
)
|
)
|
||||||
from awx.main.utils.filters import SmartFilter
|
from awx.main.utils.filters import SmartFilter
|
||||||
from awx.main.utils.named_url_graph import reset_counters
|
from awx.main.utils.named_url_graph import reset_counters
|
||||||
from awx.main.scheduler.task_manager_models import TaskManagerInstanceGroups, TaskManagerInstances
|
from awx.main.scheduler.task_manager_models import TaskManagerModels
|
||||||
from awx.main.redact import UriCleaner, REPLACE_STR
|
from awx.main.redact import UriCleaner, REPLACE_STR
|
||||||
|
|
||||||
from awx.main.validators import vars_validate_or_raise
|
from awx.main.validators import vars_validate_or_raise
|
||||||
@@ -5040,12 +5040,10 @@ class InstanceHealthCheckSerializer(BaseSerializer):
|
|||||||
class InstanceGroupSerializer(BaseSerializer):
|
class InstanceGroupSerializer(BaseSerializer):
|
||||||
|
|
||||||
show_capabilities = ['edit', 'delete']
|
show_capabilities = ['edit', 'delete']
|
||||||
|
capacity = serializers.SerializerMethodField()
|
||||||
consumed_capacity = serializers.SerializerMethodField()
|
consumed_capacity = serializers.SerializerMethodField()
|
||||||
percent_capacity_remaining = serializers.SerializerMethodField()
|
percent_capacity_remaining = serializers.SerializerMethodField()
|
||||||
jobs_running = serializers.IntegerField(
|
jobs_running = serializers.SerializerMethodField()
|
||||||
help_text=_('Count of jobs in the running or waiting state that ' 'are targeted for this instance group'), read_only=True
|
|
||||||
)
|
|
||||||
jobs_total = serializers.IntegerField(help_text=_('Count of all jobs that target this instance group'), read_only=True)
|
jobs_total = serializers.IntegerField(help_text=_('Count of all jobs that target this instance group'), read_only=True)
|
||||||
instances = serializers.SerializerMethodField()
|
instances = serializers.SerializerMethodField()
|
||||||
is_container_group = serializers.BooleanField(
|
is_container_group = serializers.BooleanField(
|
||||||
@@ -5071,6 +5069,22 @@ class InstanceGroupSerializer(BaseSerializer):
|
|||||||
label=_('Policy Instance Minimum'),
|
label=_('Policy Instance Minimum'),
|
||||||
help_text=_("Static minimum number of Instances that will be automatically assign to " "this group when new instances come online."),
|
help_text=_("Static minimum number of Instances that will be automatically assign to " "this group when new instances come online."),
|
||||||
)
|
)
|
||||||
|
max_concurrent_jobs = serializers.IntegerField(
|
||||||
|
default=0,
|
||||||
|
min_value=0,
|
||||||
|
required=False,
|
||||||
|
initial=0,
|
||||||
|
label=_('Max Concurrent Jobs'),
|
||||||
|
help_text=_("Maximum number of concurrent jobs to run on a group. When set to zero, no maximum is enforced."),
|
||||||
|
)
|
||||||
|
max_forks = serializers.IntegerField(
|
||||||
|
default=0,
|
||||||
|
min_value=0,
|
||||||
|
required=False,
|
||||||
|
initial=0,
|
||||||
|
label=_('Max Forks'),
|
||||||
|
help_text=_("Maximum number of forks to execute concurrently on a group. When set to zero, no maximum is enforced."),
|
||||||
|
)
|
||||||
policy_instance_list = serializers.ListField(
|
policy_instance_list = serializers.ListField(
|
||||||
child=serializers.CharField(),
|
child=serializers.CharField(),
|
||||||
required=False,
|
required=False,
|
||||||
@@ -5092,6 +5106,8 @@ class InstanceGroupSerializer(BaseSerializer):
|
|||||||
"consumed_capacity",
|
"consumed_capacity",
|
||||||
"percent_capacity_remaining",
|
"percent_capacity_remaining",
|
||||||
"jobs_running",
|
"jobs_running",
|
||||||
|
"max_concurrent_jobs",
|
||||||
|
"max_forks",
|
||||||
"jobs_total",
|
"jobs_total",
|
||||||
"instances",
|
"instances",
|
||||||
"is_container_group",
|
"is_container_group",
|
||||||
@@ -5173,28 +5189,39 @@ class InstanceGroupSerializer(BaseSerializer):
|
|||||||
# Store capacity values (globally computed) in the context
|
# Store capacity values (globally computed) in the context
|
||||||
if 'task_manager_igs' not in self.context:
|
if 'task_manager_igs' not in self.context:
|
||||||
instance_groups_queryset = None
|
instance_groups_queryset = None
|
||||||
jobs_qs = UnifiedJob.objects.filter(status__in=('running', 'waiting'))
|
|
||||||
if self.parent: # Is ListView:
|
if self.parent: # Is ListView:
|
||||||
instance_groups_queryset = self.parent.instance
|
instance_groups_queryset = self.parent.instance
|
||||||
|
|
||||||
instances = TaskManagerInstances(jobs_qs)
|
tm_models = TaskManagerModels.init_with_consumed_capacity(
|
||||||
instance_groups = TaskManagerInstanceGroups(instances_by_hostname=instances, instance_groups_queryset=instance_groups_queryset)
|
instance_fields=['uuid', 'version', 'capacity', 'cpu', 'memory', 'managed_by_policy', 'enabled'],
|
||||||
|
instance_groups_queryset=instance_groups_queryset,
|
||||||
|
)
|
||||||
|
|
||||||
self.context['task_manager_igs'] = instance_groups
|
self.context['task_manager_igs'] = tm_models.instance_groups
|
||||||
return self.context['task_manager_igs']
|
return self.context['task_manager_igs']
|
||||||
|
|
||||||
def get_consumed_capacity(self, obj):
|
def get_consumed_capacity(self, obj):
|
||||||
ig_mgr = self.get_ig_mgr()
|
ig_mgr = self.get_ig_mgr()
|
||||||
return ig_mgr.get_consumed_capacity(obj.name)
|
return ig_mgr.get_consumed_capacity(obj.name)
|
||||||
|
|
||||||
def get_percent_capacity_remaining(self, obj):
|
def get_capacity(self, obj):
|
||||||
if not obj.capacity:
|
|
||||||
return 0.0
|
|
||||||
ig_mgr = self.get_ig_mgr()
|
ig_mgr = self.get_ig_mgr()
|
||||||
return float("{0:.2f}".format((float(ig_mgr.get_remaining_capacity(obj.name)) / (float(obj.capacity))) * 100))
|
return ig_mgr.get_capacity(obj.name)
|
||||||
|
|
||||||
|
def get_percent_capacity_remaining(self, obj):
|
||||||
|
capacity = self.get_capacity(obj)
|
||||||
|
if not capacity:
|
||||||
|
return 0.0
|
||||||
|
consumed_capacity = self.get_consumed_capacity(obj)
|
||||||
|
return float("{0:.2f}".format(((float(capacity) - float(consumed_capacity)) / (float(capacity))) * 100))
|
||||||
|
|
||||||
def get_instances(self, obj):
|
def get_instances(self, obj):
|
||||||
return obj.instances.count()
|
ig_mgr = self.get_ig_mgr()
|
||||||
|
return len(ig_mgr.get_instances(obj.name))
|
||||||
|
|
||||||
|
def get_jobs_running(self, obj):
|
||||||
|
ig_mgr = self.get_ig_mgr()
|
||||||
|
return ig_mgr.get_jobs_running(obj.name)
|
||||||
|
|
||||||
|
|
||||||
class ActivityStreamSerializer(BaseSerializer):
|
class ActivityStreamSerializer(BaseSerializer):
|
||||||
|
|||||||
@@ -7,7 +7,7 @@ receptor_work_commands:
|
|||||||
command: ansible-runner
|
command: ansible-runner
|
||||||
params: worker
|
params: worker
|
||||||
allowruntimeparams: true
|
allowruntimeparams: true
|
||||||
verifysignature: {{ sign_work }}
|
verifysignature: true
|
||||||
custom_worksign_public_keyfile: receptor/work-public-key.pem
|
custom_worksign_public_keyfile: receptor/work-public-key.pem
|
||||||
custom_tls_certfile: receptor/tls/receptor.crt
|
custom_tls_certfile: receptor/tls/receptor.crt
|
||||||
custom_tls_keyfile: receptor/tls/receptor.key
|
custom_tls_keyfile: receptor/tls/receptor.key
|
||||||
|
|||||||
@@ -344,6 +344,13 @@ class InstanceDetail(RetrieveUpdateAPIView):
|
|||||||
model = models.Instance
|
model = models.Instance
|
||||||
serializer_class = serializers.InstanceSerializer
|
serializer_class = serializers.InstanceSerializer
|
||||||
|
|
||||||
|
def update_raw_data(self, data):
|
||||||
|
# these fields are only valid on creation of an instance, so they unwanted on detail view
|
||||||
|
data.pop('listener_port', None)
|
||||||
|
data.pop('node_type', None)
|
||||||
|
data.pop('hostname', None)
|
||||||
|
return super(InstanceDetail, self).update_raw_data(data)
|
||||||
|
|
||||||
def update(self, request, *args, **kwargs):
|
def update(self, request, *args, **kwargs):
|
||||||
r = super(InstanceDetail, self).update(request, *args, **kwargs)
|
r = super(InstanceDetail, self).update(request, *args, **kwargs)
|
||||||
if status.is_success(r.status_code):
|
if status.is_success(r.status_code):
|
||||||
|
|||||||
@@ -16,7 +16,7 @@ from rest_framework import status
|
|||||||
|
|
||||||
from awx.main.constants import ACTIVE_STATES
|
from awx.main.constants import ACTIVE_STATES
|
||||||
from awx.main.utils import get_object_or_400
|
from awx.main.utils import get_object_or_400
|
||||||
from awx.main.models.ha import Instance, InstanceGroup
|
from awx.main.models.ha import Instance, InstanceGroup, schedule_policy_task
|
||||||
from awx.main.models.organization import Team
|
from awx.main.models.organization import Team
|
||||||
from awx.main.models.projects import Project
|
from awx.main.models.projects import Project
|
||||||
from awx.main.models.inventory import Inventory
|
from awx.main.models.inventory import Inventory
|
||||||
@@ -107,6 +107,11 @@ class InstanceGroupMembershipMixin(object):
|
|||||||
if inst_name in ig_obj.policy_instance_list:
|
if inst_name in ig_obj.policy_instance_list:
|
||||||
ig_obj.policy_instance_list.pop(ig_obj.policy_instance_list.index(inst_name))
|
ig_obj.policy_instance_list.pop(ig_obj.policy_instance_list.index(inst_name))
|
||||||
ig_obj.save(update_fields=['policy_instance_list'])
|
ig_obj.save(update_fields=['policy_instance_list'])
|
||||||
|
|
||||||
|
# sometimes removing an instance has a non-obvious consequence
|
||||||
|
# this is almost always true if policy_instance_percentage or _minimum is non-zero
|
||||||
|
# after removing a single instance, the other memberships need to be re-balanced
|
||||||
|
schedule_policy_task()
|
||||||
return response
|
return response
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
@@ -2697,46 +2697,66 @@ class ActivityStreamAccess(BaseAccess):
|
|||||||
# 'job_template', 'job', 'project', 'project_update', 'workflow_job',
|
# 'job_template', 'job', 'project', 'project_update', 'workflow_job',
|
||||||
# 'inventory_source', 'workflow_job_template'
|
# 'inventory_source', 'workflow_job_template'
|
||||||
|
|
||||||
inventory_set = Inventory.accessible_objects(self.user, 'read_role')
|
q = Q(user=self.user)
|
||||||
credential_set = Credential.accessible_objects(self.user, 'read_role')
|
inventory_set = Inventory.accessible_pk_qs(self.user, 'read_role')
|
||||||
auditing_orgs = (
|
if inventory_set:
|
||||||
(Organization.accessible_objects(self.user, 'admin_role') | Organization.accessible_objects(self.user, 'auditor_role'))
|
q |= (
|
||||||
.distinct()
|
|
||||||
.values_list('id', flat=True)
|
|
||||||
)
|
|
||||||
project_set = Project.accessible_objects(self.user, 'read_role')
|
|
||||||
jt_set = JobTemplate.accessible_objects(self.user, 'read_role')
|
|
||||||
team_set = Team.accessible_objects(self.user, 'read_role')
|
|
||||||
wfjt_set = WorkflowJobTemplate.accessible_objects(self.user, 'read_role')
|
|
||||||
app_set = OAuth2ApplicationAccess(self.user).filtered_queryset()
|
|
||||||
token_set = OAuth2TokenAccess(self.user).filtered_queryset()
|
|
||||||
|
|
||||||
return qs.filter(
|
|
||||||
Q(ad_hoc_command__inventory__in=inventory_set)
|
Q(ad_hoc_command__inventory__in=inventory_set)
|
||||||
| Q(o_auth2_application__in=app_set)
|
|
||||||
| Q(o_auth2_access_token__in=token_set)
|
|
||||||
| Q(user__in=auditing_orgs.values('member_role__members'))
|
|
||||||
| Q(user=self.user)
|
|
||||||
| Q(organization__in=auditing_orgs)
|
|
||||||
| Q(inventory__in=inventory_set)
|
| Q(inventory__in=inventory_set)
|
||||||
| Q(host__inventory__in=inventory_set)
|
| Q(host__inventory__in=inventory_set)
|
||||||
| Q(group__inventory__in=inventory_set)
|
| Q(group__inventory__in=inventory_set)
|
||||||
| Q(inventory_source__inventory__in=inventory_set)
|
| Q(inventory_source__inventory__in=inventory_set)
|
||||||
| Q(inventory_update__inventory_source__inventory__in=inventory_set)
|
| Q(inventory_update__inventory_source__inventory__in=inventory_set)
|
||||||
| Q(credential__in=credential_set)
|
)
|
||||||
| Q(team__in=team_set)
|
|
||||||
| Q(project__in=project_set)
|
credential_set = Credential.accessible_pk_qs(self.user, 'read_role')
|
||||||
| Q(project_update__project__in=project_set)
|
if credential_set:
|
||||||
| Q(job_template__in=jt_set)
|
q |= Q(credential__in=credential_set)
|
||||||
| Q(job__job_template__in=jt_set)
|
|
||||||
| Q(workflow_job_template__in=wfjt_set)
|
auditing_orgs = (
|
||||||
| Q(workflow_job_template_node__workflow_job_template__in=wfjt_set)
|
(Organization.accessible_objects(self.user, 'admin_role') | Organization.accessible_objects(self.user, 'auditor_role'))
|
||||||
| Q(workflow_job__workflow_job_template__in=wfjt_set)
|
.distinct()
|
||||||
|
.values_list('id', flat=True)
|
||||||
|
)
|
||||||
|
if auditing_orgs:
|
||||||
|
q |= (
|
||||||
|
Q(user__in=auditing_orgs.values('member_role__members'))
|
||||||
|
| Q(organization__in=auditing_orgs)
|
||||||
| Q(notification_template__organization__in=auditing_orgs)
|
| Q(notification_template__organization__in=auditing_orgs)
|
||||||
| Q(notification__notification_template__organization__in=auditing_orgs)
|
| Q(notification__notification_template__organization__in=auditing_orgs)
|
||||||
| Q(label__organization__in=auditing_orgs)
|
| Q(label__organization__in=auditing_orgs)
|
||||||
| Q(role__in=Role.objects.filter(ancestors__in=self.user.roles.all()) if auditing_orgs else [])
|
| Q(role__in=Role.objects.filter(ancestors__in=self.user.roles.all()) if auditing_orgs else [])
|
||||||
).distinct()
|
)
|
||||||
|
|
||||||
|
project_set = Project.accessible_pk_qs(self.user, 'read_role')
|
||||||
|
if project_set:
|
||||||
|
q |= Q(project__in=project_set) | Q(project_update__project__in=project_set)
|
||||||
|
|
||||||
|
jt_set = JobTemplate.accessible_pk_qs(self.user, 'read_role')
|
||||||
|
if jt_set:
|
||||||
|
q |= Q(job_template__in=jt_set) | Q(job__job_template__in=jt_set)
|
||||||
|
|
||||||
|
wfjt_set = WorkflowJobTemplate.accessible_pk_qs(self.user, 'read_role')
|
||||||
|
if wfjt_set:
|
||||||
|
q |= (
|
||||||
|
Q(workflow_job_template__in=wfjt_set)
|
||||||
|
| Q(workflow_job_template_node__workflow_job_template__in=wfjt_set)
|
||||||
|
| Q(workflow_job__workflow_job_template__in=wfjt_set)
|
||||||
|
)
|
||||||
|
|
||||||
|
team_set = Team.accessible_pk_qs(self.user, 'read_role')
|
||||||
|
if team_set:
|
||||||
|
q |= Q(team__in=team_set)
|
||||||
|
|
||||||
|
app_set = OAuth2ApplicationAccess(self.user).filtered_queryset()
|
||||||
|
if app_set:
|
||||||
|
q |= Q(o_auth2_application__in=app_set)
|
||||||
|
|
||||||
|
token_set = OAuth2TokenAccess(self.user).filtered_queryset()
|
||||||
|
if token_set:
|
||||||
|
q |= Q(o_auth2_access_token__in=token_set)
|
||||||
|
|
||||||
|
return qs.filter(q).distinct()
|
||||||
|
|
||||||
def can_add(self, data):
|
def can_add(self, data):
|
||||||
return False
|
return False
|
||||||
|
|||||||
@@ -1,8 +1,8 @@
|
|||||||
import datetime
|
import datetime
|
||||||
import asyncio
|
import asyncio
|
||||||
import logging
|
import logging
|
||||||
import aioredis
|
|
||||||
import redis
|
import redis
|
||||||
|
import redis.asyncio
|
||||||
import re
|
import re
|
||||||
|
|
||||||
from prometheus_client import (
|
from prometheus_client import (
|
||||||
@@ -82,7 +82,7 @@ class BroadcastWebsocketStatsManager:
|
|||||||
|
|
||||||
async def run_loop(self):
|
async def run_loop(self):
|
||||||
try:
|
try:
|
||||||
redis_conn = await aioredis.create_redis_pool(settings.BROKER_URL)
|
redis_conn = await redis.asyncio.Redis.from_url(settings.BROKER_URL)
|
||||||
while True:
|
while True:
|
||||||
stats_data_str = ''.join(stat.serialize() for stat in self._stats.values())
|
stats_data_str = ''.join(stat.serialize() for stat in self._stats.values())
|
||||||
await redis_conn.set(self._redis_key, stats_data_str)
|
await redis_conn.set(self._redis_key, stats_data_str)
|
||||||
@@ -122,8 +122,8 @@ class BroadcastWebsocketStats:
|
|||||||
'Number of messages received, to be forwarded, by the broadcast websocket system',
|
'Number of messages received, to be forwarded, by the broadcast websocket system',
|
||||||
registry=self._registry,
|
registry=self._registry,
|
||||||
)
|
)
|
||||||
self._messages_received = Gauge(
|
self._messages_received_current_conn = Gauge(
|
||||||
f'awx_{self.remote_name}_messages_received',
|
f'awx_{self.remote_name}_messages_received_currrent_conn',
|
||||||
'Number forwarded messages received by the broadcast websocket system, for the duration of the current connection',
|
'Number forwarded messages received by the broadcast websocket system, for the duration of the current connection',
|
||||||
registry=self._registry,
|
registry=self._registry,
|
||||||
)
|
)
|
||||||
@@ -144,13 +144,13 @@ class BroadcastWebsocketStats:
|
|||||||
|
|
||||||
def record_message_received(self):
|
def record_message_received(self):
|
||||||
self._internal_messages_received_per_minute.record()
|
self._internal_messages_received_per_minute.record()
|
||||||
self._messages_received.inc()
|
self._messages_received_current_conn.inc()
|
||||||
self._messages_received_total.inc()
|
self._messages_received_total.inc()
|
||||||
|
|
||||||
def record_connection_established(self):
|
def record_connection_established(self):
|
||||||
self._connection.state('connected')
|
self._connection.state('connected')
|
||||||
self._connection_start.set_to_current_time()
|
self._connection_start.set_to_current_time()
|
||||||
self._messages_received.set(0)
|
self._messages_received_current_conn.set(0)
|
||||||
|
|
||||||
def record_connection_lost(self):
|
def record_connection_lost(self):
|
||||||
self._connection.state('disconnected')
|
self._connection.state('disconnected')
|
||||||
|
|||||||
@@ -16,7 +16,7 @@ from awx.conf.license import get_license
|
|||||||
from awx.main.utils import get_awx_version, camelcase_to_underscore, datetime_hook
|
from awx.main.utils import get_awx_version, camelcase_to_underscore, datetime_hook
|
||||||
from awx.main import models
|
from awx.main import models
|
||||||
from awx.main.analytics import register
|
from awx.main.analytics import register
|
||||||
from awx.main.scheduler.task_manager_models import TaskManagerInstances
|
from awx.main.scheduler.task_manager_models import TaskManagerModels
|
||||||
|
|
||||||
"""
|
"""
|
||||||
This module is used to define metrics collected by awx.main.analytics.gather()
|
This module is used to define metrics collected by awx.main.analytics.gather()
|
||||||
@@ -237,11 +237,8 @@ def projects_by_scm_type(since, **kwargs):
|
|||||||
def instance_info(since, include_hostnames=False, **kwargs):
|
def instance_info(since, include_hostnames=False, **kwargs):
|
||||||
info = {}
|
info = {}
|
||||||
# Use same method that the TaskManager does to compute consumed capacity without querying all running jobs for each Instance
|
# Use same method that the TaskManager does to compute consumed capacity without querying all running jobs for each Instance
|
||||||
active_tasks = models.UnifiedJob.objects.filter(status__in=['running', 'waiting']).only('task_impact', 'controller_node', 'execution_node')
|
tm_models = TaskManagerModels.init_with_consumed_capacity(instance_fields=['uuid', 'version', 'capacity', 'cpu', 'memory', 'managed_by_policy', 'enabled'])
|
||||||
tm_instances = TaskManagerInstances(
|
for tm_instance in tm_models.instances.instances_by_hostname.values():
|
||||||
active_tasks, instance_fields=['uuid', 'version', 'capacity', 'cpu', 'memory', 'managed_by_policy', 'enabled', 'node_type']
|
|
||||||
)
|
|
||||||
for tm_instance in tm_instances.instances_by_hostname.values():
|
|
||||||
instance = tm_instance.obj
|
instance = tm_instance.obj
|
||||||
instance_info = {
|
instance_info = {
|
||||||
'uuid': instance.uuid,
|
'uuid': instance.uuid,
|
||||||
|
|||||||
@@ -569,7 +569,7 @@ register(
|
|||||||
register(
|
register(
|
||||||
'LOG_AGGREGATOR_LOGGERS',
|
'LOG_AGGREGATOR_LOGGERS',
|
||||||
field_class=fields.StringListField,
|
field_class=fields.StringListField,
|
||||||
default=['awx', 'activity_stream', 'job_events', 'system_tracking'],
|
default=['awx', 'activity_stream', 'job_events', 'system_tracking', 'broadcast_websocket'],
|
||||||
label=_('Loggers Sending Data to Log Aggregator Form'),
|
label=_('Loggers Sending Data to Log Aggregator Form'),
|
||||||
help_text=_(
|
help_text=_(
|
||||||
'List of loggers that will send HTTP logs to the collector, these can '
|
'List of loggers that will send HTTP logs to the collector, these can '
|
||||||
@@ -577,7 +577,8 @@ register(
|
|||||||
'awx - service logs\n'
|
'awx - service logs\n'
|
||||||
'activity_stream - activity stream records\n'
|
'activity_stream - activity stream records\n'
|
||||||
'job_events - callback data from Ansible job events\n'
|
'job_events - callback data from Ansible job events\n'
|
||||||
'system_tracking - facts gathered from scan jobs.'
|
'system_tracking - facts gathered from scan jobs\n'
|
||||||
|
'broadcast_websocket - errors pertaining to websockets broadcast metrics\n'
|
||||||
),
|
),
|
||||||
category=_('Logging'),
|
category=_('Logging'),
|
||||||
category_slug='logging',
|
category_slug='logging',
|
||||||
|
|||||||
@@ -9,10 +9,16 @@ aim_inputs = {
|
|||||||
'fields': [
|
'fields': [
|
||||||
{
|
{
|
||||||
'id': 'url',
|
'id': 'url',
|
||||||
'label': _('CyberArk AIM URL'),
|
'label': _('CyberArk CCP URL'),
|
||||||
'type': 'string',
|
'type': 'string',
|
||||||
'format': 'url',
|
'format': 'url',
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
'id': 'webservice_id',
|
||||||
|
'label': _('Web Service ID'),
|
||||||
|
'type': 'string',
|
||||||
|
'help_text': _('The CCP Web Service ID. Leave blank to default to AIMWebService.'),
|
||||||
|
},
|
||||||
{
|
{
|
||||||
'id': 'app_id',
|
'id': 'app_id',
|
||||||
'label': _('Application ID'),
|
'label': _('Application ID'),
|
||||||
@@ -64,10 +70,13 @@ def aim_backend(**kwargs):
|
|||||||
client_cert = kwargs.get('client_cert', None)
|
client_cert = kwargs.get('client_cert', None)
|
||||||
client_key = kwargs.get('client_key', None)
|
client_key = kwargs.get('client_key', None)
|
||||||
verify = kwargs['verify']
|
verify = kwargs['verify']
|
||||||
|
webservice_id = kwargs['webservice_id']
|
||||||
app_id = kwargs['app_id']
|
app_id = kwargs['app_id']
|
||||||
object_query = kwargs['object_query']
|
object_query = kwargs['object_query']
|
||||||
object_query_format = kwargs['object_query_format']
|
object_query_format = kwargs['object_query_format']
|
||||||
reason = kwargs.get('reason', None)
|
reason = kwargs.get('reason', None)
|
||||||
|
if webservice_id == '':
|
||||||
|
webservice_id = 'AIMWebService'
|
||||||
|
|
||||||
query_params = {
|
query_params = {
|
||||||
'AppId': app_id,
|
'AppId': app_id,
|
||||||
@@ -78,7 +87,7 @@ def aim_backend(**kwargs):
|
|||||||
query_params['reason'] = reason
|
query_params['reason'] = reason
|
||||||
|
|
||||||
request_qs = '?' + urlencode(query_params, quote_via=quote)
|
request_qs = '?' + urlencode(query_params, quote_via=quote)
|
||||||
request_url = urljoin(url, '/'.join(['AIMWebService', 'api', 'Accounts']))
|
request_url = urljoin(url, '/'.join([webservice_id, 'api', 'Accounts']))
|
||||||
|
|
||||||
with CertFiles(client_cert, client_key) as cert:
|
with CertFiles(client_cert, client_key) as cert:
|
||||||
res = requests.get(
|
res = requests.get(
|
||||||
@@ -92,4 +101,4 @@ def aim_backend(**kwargs):
|
|||||||
return res.json()['Content']
|
return res.json()['Content']
|
||||||
|
|
||||||
|
|
||||||
aim_plugin = CredentialPlugin('CyberArk AIM Central Credential Provider Lookup', inputs=aim_inputs, backend=aim_backend)
|
aim_plugin = CredentialPlugin('CyberArk Central Credential Provider Lookup', inputs=aim_inputs, backend=aim_backend)
|
||||||
|
|||||||
@@ -38,7 +38,14 @@ class Command(BaseCommand):
|
|||||||
(changed, instance) = Instance.objects.register(ip_address=os.environ.get('MY_POD_IP'), node_type='control', uuid=settings.SYSTEM_UUID)
|
(changed, instance) = Instance.objects.register(ip_address=os.environ.get('MY_POD_IP'), node_type='control', uuid=settings.SYSTEM_UUID)
|
||||||
RegisterQueue(settings.DEFAULT_CONTROL_PLANE_QUEUE_NAME, 100, 0, [], is_container_group=False).register()
|
RegisterQueue(settings.DEFAULT_CONTROL_PLANE_QUEUE_NAME, 100, 0, [], is_container_group=False).register()
|
||||||
RegisterQueue(
|
RegisterQueue(
|
||||||
settings.DEFAULT_EXECUTION_QUEUE_NAME, 100, 0, [], is_container_group=True, pod_spec_override=settings.DEFAULT_EXECUTION_QUEUE_POD_SPEC_OVERRIDE
|
settings.DEFAULT_EXECUTION_QUEUE_NAME,
|
||||||
|
100,
|
||||||
|
0,
|
||||||
|
[],
|
||||||
|
is_container_group=True,
|
||||||
|
pod_spec_override=settings.DEFAULT_EXECUTION_QUEUE_POD_SPEC_OVERRIDE,
|
||||||
|
max_forks=settings.DEFAULT_EXECUTION_QUEUE_MAX_FORKS,
|
||||||
|
max_concurrent_jobs=settings.DEFAULT_EXECUTION_QUEUE_MAX_CONCURRENT_JOBS,
|
||||||
).register()
|
).register()
|
||||||
else:
|
else:
|
||||||
(changed, instance) = Instance.objects.register(hostname=hostname, node_type=node_type, uuid=uuid)
|
(changed, instance) = Instance.objects.register(hostname=hostname, node_type=node_type, uuid=uuid)
|
||||||
|
|||||||
@@ -32,8 +32,14 @@ class Command(BaseCommand):
|
|||||||
def handle(self, **options):
|
def handle(self, **options):
|
||||||
self.old_key = settings.SECRET_KEY
|
self.old_key = settings.SECRET_KEY
|
||||||
custom_key = os.environ.get("TOWER_SECRET_KEY")
|
custom_key = os.environ.get("TOWER_SECRET_KEY")
|
||||||
if options.get("use_custom_key") and custom_key:
|
if options.get("use_custom_key"):
|
||||||
|
if custom_key:
|
||||||
self.new_key = custom_key
|
self.new_key = custom_key
|
||||||
|
else:
|
||||||
|
print("Use custom key was specified but the env var TOWER_SECRET_KEY was not available")
|
||||||
|
import sys
|
||||||
|
|
||||||
|
sys.exit(1)
|
||||||
else:
|
else:
|
||||||
self.new_key = base64.encodebytes(os.urandom(33)).decode().rstrip()
|
self.new_key = base64.encodebytes(os.urandom(33)).decode().rstrip()
|
||||||
self._notification_templates()
|
self._notification_templates()
|
||||||
|
|||||||
@@ -17,7 +17,9 @@ class InstanceNotFound(Exception):
|
|||||||
|
|
||||||
|
|
||||||
class RegisterQueue:
|
class RegisterQueue:
|
||||||
def __init__(self, queuename, instance_percent, inst_min, hostname_list, is_container_group=None, pod_spec_override=None):
|
def __init__(
|
||||||
|
self, queuename, instance_percent, inst_min, hostname_list, is_container_group=None, pod_spec_override=None, max_forks=None, max_concurrent_jobs=None
|
||||||
|
):
|
||||||
self.instance_not_found_err = None
|
self.instance_not_found_err = None
|
||||||
self.queuename = queuename
|
self.queuename = queuename
|
||||||
self.instance_percent = instance_percent
|
self.instance_percent = instance_percent
|
||||||
@@ -25,6 +27,8 @@ class RegisterQueue:
|
|||||||
self.hostname_list = hostname_list
|
self.hostname_list = hostname_list
|
||||||
self.is_container_group = is_container_group
|
self.is_container_group = is_container_group
|
||||||
self.pod_spec_override = pod_spec_override
|
self.pod_spec_override = pod_spec_override
|
||||||
|
self.max_forks = max_forks
|
||||||
|
self.max_concurrent_jobs = max_concurrent_jobs
|
||||||
|
|
||||||
def get_create_update_instance_group(self):
|
def get_create_update_instance_group(self):
|
||||||
created = False
|
created = False
|
||||||
@@ -45,6 +49,14 @@ class RegisterQueue:
|
|||||||
ig.pod_spec_override = self.pod_spec_override
|
ig.pod_spec_override = self.pod_spec_override
|
||||||
changed = True
|
changed = True
|
||||||
|
|
||||||
|
if self.max_forks and (ig.max_forks != self.max_forks):
|
||||||
|
ig.max_forks = self.max_forks
|
||||||
|
changed = True
|
||||||
|
|
||||||
|
if self.max_concurrent_jobs and (ig.max_concurrent_jobs != self.max_concurrent_jobs):
|
||||||
|
ig.max_concurrent_jobs = self.max_concurrent_jobs
|
||||||
|
changed = True
|
||||||
|
|
||||||
if changed:
|
if changed:
|
||||||
ig.save()
|
ig.save()
|
||||||
|
|
||||||
|
|||||||
@@ -158,7 +158,11 @@ class InstanceManager(models.Manager):
|
|||||||
return (False, instance)
|
return (False, instance)
|
||||||
|
|
||||||
# Create new instance, and fill in default values
|
# Create new instance, and fill in default values
|
||||||
create_defaults = {'node_state': Instance.States.INSTALLED, 'capacity': 0}
|
create_defaults = {
|
||||||
|
'node_state': Instance.States.INSTALLED,
|
||||||
|
'capacity': 0,
|
||||||
|
'listener_port': 27199,
|
||||||
|
}
|
||||||
if defaults is not None:
|
if defaults is not None:
|
||||||
create_defaults.update(defaults)
|
create_defaults.update(defaults)
|
||||||
uuid_option = {}
|
uuid_option = {}
|
||||||
|
|||||||
@@ -1,24 +1,14 @@
|
|||||||
# Generated by Django 3.2.13 on 2022-06-21 21:29
|
# Generated by Django 3.2.13 on 2022-06-21 21:29
|
||||||
|
|
||||||
from django.db import migrations
|
from django.db import migrations
|
||||||
import logging
|
|
||||||
|
|
||||||
logger = logging.getLogger("awx")
|
|
||||||
|
|
||||||
|
|
||||||
def forwards(apps, schema_editor):
|
def forwards(apps, schema_editor):
|
||||||
InventorySource = apps.get_model('main', 'InventorySource')
|
InventorySource = apps.get_model('main', 'InventorySource')
|
||||||
sources = InventorySource.objects.filter(update_on_project_update=True)
|
InventorySource.objects.filter(update_on_project_update=True).update(update_on_launch=True)
|
||||||
for src in sources:
|
|
||||||
if src.update_on_launch == False:
|
Project = apps.get_model('main', 'Project')
|
||||||
src.update_on_launch = True
|
Project.objects.filter(scm_inventory_sources__update_on_project_update=True).update(scm_update_on_launch=True)
|
||||||
src.save(update_fields=['update_on_launch'])
|
|
||||||
logger.info(f"Setting update_on_launch to True for {src}")
|
|
||||||
proj = src.source_project
|
|
||||||
if proj and proj.scm_update_on_launch is False:
|
|
||||||
proj.scm_update_on_launch = True
|
|
||||||
proj.save(update_fields=['scm_update_on_launch'])
|
|
||||||
logger.warning(f"Setting scm_update_on_launch to True for {proj}")
|
|
||||||
|
|
||||||
|
|
||||||
class Migration(migrations.Migration):
|
class Migration(migrations.Migration):
|
||||||
|
|||||||
23
awx/main/migrations/0173_instancegroup_max_limits.py
Normal file
23
awx/main/migrations/0173_instancegroup_max_limits.py
Normal file
@@ -0,0 +1,23 @@
|
|||||||
|
# Generated by Django 3.2.13 on 2022-10-24 18:22
|
||||||
|
|
||||||
|
from django.db import migrations, models
|
||||||
|
|
||||||
|
|
||||||
|
class Migration(migrations.Migration):
|
||||||
|
|
||||||
|
dependencies = [
|
||||||
|
('main', '0172_prevent_instance_fallback'),
|
||||||
|
]
|
||||||
|
|
||||||
|
operations = [
|
||||||
|
migrations.AddField(
|
||||||
|
model_name='instancegroup',
|
||||||
|
name='max_concurrent_jobs',
|
||||||
|
field=models.IntegerField(default=0, help_text='Maximum number of concurrent jobs to run on this group. Zero means no limit.'),
|
||||||
|
),
|
||||||
|
migrations.AddField(
|
||||||
|
model_name='instancegroup',
|
||||||
|
name='max_forks',
|
||||||
|
field=models.IntegerField(default=0, help_text='Max forks to execute on this group. Zero means no limit.'),
|
||||||
|
),
|
||||||
|
]
|
||||||
18
awx/main/migrations/0174_ensure_org_ee_admin_roles.py
Normal file
18
awx/main/migrations/0174_ensure_org_ee_admin_roles.py
Normal file
@@ -0,0 +1,18 @@
|
|||||||
|
# Generated by Django 3.2.16 on 2022-12-07 21:11
|
||||||
|
|
||||||
|
from django.db import migrations
|
||||||
|
|
||||||
|
from awx.main.migrations import _rbac as rbac
|
||||||
|
from awx.main.migrations import _migration_utils as migration_utils
|
||||||
|
|
||||||
|
|
||||||
|
class Migration(migrations.Migration):
|
||||||
|
|
||||||
|
dependencies = [
|
||||||
|
('main', '0173_instancegroup_max_limits'),
|
||||||
|
]
|
||||||
|
|
||||||
|
operations = [
|
||||||
|
migrations.RunPython(migration_utils.set_current_apps_for_migrations),
|
||||||
|
migrations.RunPython(rbac.create_roles),
|
||||||
|
]
|
||||||
@@ -15,6 +15,7 @@ def aws(cred, env, private_data_dir):
|
|||||||
|
|
||||||
if cred.has_input('security_token'):
|
if cred.has_input('security_token'):
|
||||||
env['AWS_SECURITY_TOKEN'] = cred.get_input('security_token', default='')
|
env['AWS_SECURITY_TOKEN'] = cred.get_input('security_token', default='')
|
||||||
|
env['AWS_SESSION_TOKEN'] = env['AWS_SECURITY_TOKEN']
|
||||||
|
|
||||||
|
|
||||||
def gce(cred, env, private_data_dir):
|
def gce(cred, env, private_data_dir):
|
||||||
|
|||||||
@@ -379,6 +379,8 @@ class InstanceGroup(HasPolicyEditsMixin, BaseModel, RelatedJobsMixin):
|
|||||||
default='',
|
default='',
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
|
max_concurrent_jobs = models.IntegerField(default=0, help_text=_("Maximum number of concurrent jobs to run on this group. Zero means no limit."))
|
||||||
|
max_forks = models.IntegerField(default=0, help_text=_("Max forks to execute on this group. Zero means no limit."))
|
||||||
policy_instance_percentage = models.IntegerField(default=0, help_text=_("Percentage of Instances to automatically assign to this group"))
|
policy_instance_percentage = models.IntegerField(default=0, help_text=_("Percentage of Instances to automatically assign to this group"))
|
||||||
policy_instance_minimum = models.IntegerField(default=0, help_text=_("Static minimum number of Instances to automatically assign to this group"))
|
policy_instance_minimum = models.IntegerField(default=0, help_text=_("Static minimum number of Instances to automatically assign to this group"))
|
||||||
policy_instance_list = JSONBlob(
|
policy_instance_list = JSONBlob(
|
||||||
@@ -392,6 +394,8 @@ class InstanceGroup(HasPolicyEditsMixin, BaseModel, RelatedJobsMixin):
|
|||||||
|
|
||||||
@property
|
@property
|
||||||
def capacity(self):
|
def capacity(self):
|
||||||
|
if self.is_container_group:
|
||||||
|
return self.max_forks
|
||||||
return sum(inst.capacity for inst in self.instances.all())
|
return sum(inst.capacity for inst in self.instances.all())
|
||||||
|
|
||||||
@property
|
@property
|
||||||
|
|||||||
@@ -27,8 +27,8 @@ class AWXProtocolTypeRouter(ProtocolTypeRouter):
|
|||||||
|
|
||||||
|
|
||||||
websocket_urlpatterns = [
|
websocket_urlpatterns = [
|
||||||
re_path(r'websocket/$', consumers.EventConsumer),
|
re_path(r'websocket/$', consumers.EventConsumer.as_asgi()),
|
||||||
re_path(r'websocket/broadcast/$', consumers.BroadcastConsumer),
|
re_path(r'websocket/broadcast/$', consumers.BroadcastConsumer.as_asgi()),
|
||||||
]
|
]
|
||||||
|
|
||||||
application = AWXProtocolTypeRouter(
|
application = AWXProtocolTypeRouter(
|
||||||
|
|||||||
@@ -43,8 +43,7 @@ from awx.main.utils.common import task_manager_bulk_reschedule, is_testing
|
|||||||
from awx.main.signals import disable_activity_stream
|
from awx.main.signals import disable_activity_stream
|
||||||
from awx.main.constants import ACTIVE_STATES
|
from awx.main.constants import ACTIVE_STATES
|
||||||
from awx.main.scheduler.dependency_graph import DependencyGraph
|
from awx.main.scheduler.dependency_graph import DependencyGraph
|
||||||
from awx.main.scheduler.task_manager_models import TaskManagerInstances
|
from awx.main.scheduler.task_manager_models import TaskManagerModels
|
||||||
from awx.main.scheduler.task_manager_models import TaskManagerInstanceGroups
|
|
||||||
import awx.main.analytics.subsystem_metrics as s_metrics
|
import awx.main.analytics.subsystem_metrics as s_metrics
|
||||||
from awx.main.utils import decrypt_field
|
from awx.main.utils import decrypt_field
|
||||||
|
|
||||||
@@ -71,7 +70,12 @@ class TaskBase:
|
|||||||
# is called later.
|
# is called later.
|
||||||
self.subsystem_metrics = s_metrics.Metrics(auto_pipe_execute=False)
|
self.subsystem_metrics = s_metrics.Metrics(auto_pipe_execute=False)
|
||||||
self.start_time = time.time()
|
self.start_time = time.time()
|
||||||
|
|
||||||
|
# We want to avoid calling settings in loops, so cache these settings at init time
|
||||||
self.start_task_limit = settings.START_TASK_LIMIT
|
self.start_task_limit = settings.START_TASK_LIMIT
|
||||||
|
self.task_manager_timeout = settings.TASK_MANAGER_TIMEOUT
|
||||||
|
self.control_task_impact = settings.AWX_CONTROL_NODE_TASK_IMPACT
|
||||||
|
|
||||||
for m in self.subsystem_metrics.METRICS:
|
for m in self.subsystem_metrics.METRICS:
|
||||||
if m.startswith(self.prefix):
|
if m.startswith(self.prefix):
|
||||||
self.subsystem_metrics.set(m, 0)
|
self.subsystem_metrics.set(m, 0)
|
||||||
@@ -79,7 +83,7 @@ class TaskBase:
|
|||||||
def timed_out(self):
|
def timed_out(self):
|
||||||
"""Return True/False if we have met or exceeded the timeout for the task manager."""
|
"""Return True/False if we have met or exceeded the timeout for the task manager."""
|
||||||
elapsed = time.time() - self.start_time
|
elapsed = time.time() - self.start_time
|
||||||
if elapsed >= settings.TASK_MANAGER_TIMEOUT:
|
if elapsed >= self.task_manager_timeout:
|
||||||
logger.warning(f"{self.prefix} manager has run for {elapsed} which is greater than TASK_MANAGER_TIMEOUT of {settings.TASK_MANAGER_TIMEOUT}.")
|
logger.warning(f"{self.prefix} manager has run for {elapsed} which is greater than TASK_MANAGER_TIMEOUT of {settings.TASK_MANAGER_TIMEOUT}.")
|
||||||
return True
|
return True
|
||||||
return False
|
return False
|
||||||
@@ -471,9 +475,8 @@ class TaskManager(TaskBase):
|
|||||||
Init AFTER we know this instance of the task manager will run because the lock is acquired.
|
Init AFTER we know this instance of the task manager will run because the lock is acquired.
|
||||||
"""
|
"""
|
||||||
self.dependency_graph = DependencyGraph()
|
self.dependency_graph = DependencyGraph()
|
||||||
self.instances = TaskManagerInstances(self.all_tasks)
|
self.tm_models = TaskManagerModels()
|
||||||
self.instance_groups = TaskManagerInstanceGroups(instances_by_hostname=self.instances)
|
self.controlplane_ig = self.tm_models.instance_groups.controlplane_ig
|
||||||
self.controlplane_ig = self.instance_groups.controlplane_ig
|
|
||||||
|
|
||||||
def job_blocked_by(self, task):
|
def job_blocked_by(self, task):
|
||||||
# TODO: I'm not happy with this, I think blocking behavior should be decided outside of the dependency graph
|
# TODO: I'm not happy with this, I think blocking behavior should be decided outside of the dependency graph
|
||||||
@@ -504,8 +507,16 @@ class TaskManager(TaskBase):
|
|||||||
return None
|
return None
|
||||||
|
|
||||||
@timeit
|
@timeit
|
||||||
def start_task(self, task, instance_group, dependent_tasks=None, instance=None):
|
def start_task(self, task, instance_group, instance=None):
|
||||||
|
# Just like for process_running_tasks, add the job to the dependency graph and
|
||||||
|
# ask the TaskManagerInstanceGroups object to update consumed capacity on all
|
||||||
|
# implicated instances and container groups.
|
||||||
self.dependency_graph.add_job(task)
|
self.dependency_graph.add_job(task)
|
||||||
|
if instance_group is not None:
|
||||||
|
task.instance_group = instance_group
|
||||||
|
# We need the instance group assigned to correctly account for container group max_concurrent_jobs and max_forks
|
||||||
|
self.tm_models.consume_capacity(task)
|
||||||
|
|
||||||
self.subsystem_metrics.inc(f"{self.prefix}_tasks_started", 1)
|
self.subsystem_metrics.inc(f"{self.prefix}_tasks_started", 1)
|
||||||
self.start_task_limit -= 1
|
self.start_task_limit -= 1
|
||||||
if self.start_task_limit == 0:
|
if self.start_task_limit == 0:
|
||||||
@@ -513,20 +524,6 @@ class TaskManager(TaskBase):
|
|||||||
ScheduleTaskManager().schedule()
|
ScheduleTaskManager().schedule()
|
||||||
from awx.main.tasks.system import handle_work_error, handle_work_success
|
from awx.main.tasks.system import handle_work_error, handle_work_success
|
||||||
|
|
||||||
# update capacity for control node and execution node
|
|
||||||
if task.controller_node:
|
|
||||||
self.instances[task.controller_node].consume_capacity(settings.AWX_CONTROL_NODE_TASK_IMPACT)
|
|
||||||
if task.execution_node:
|
|
||||||
self.instances[task.execution_node].consume_capacity(task.task_impact)
|
|
||||||
|
|
||||||
dependent_tasks = dependent_tasks or []
|
|
||||||
|
|
||||||
task_actual = {
|
|
||||||
'type': get_type_for_model(type(task)),
|
|
||||||
'id': task.id,
|
|
||||||
}
|
|
||||||
dependencies = [{'type': get_type_for_model(type(t)), 'id': t.id} for t in dependent_tasks]
|
|
||||||
|
|
||||||
task.status = 'waiting'
|
task.status = 'waiting'
|
||||||
|
|
||||||
(start_status, opts) = task.pre_start()
|
(start_status, opts) = task.pre_start()
|
||||||
@@ -546,7 +543,6 @@ class TaskManager(TaskBase):
|
|||||||
ScheduleWorkflowManager().schedule()
|
ScheduleWorkflowManager().schedule()
|
||||||
# at this point we already have control/execution nodes selected for the following cases
|
# at this point we already have control/execution nodes selected for the following cases
|
||||||
else:
|
else:
|
||||||
task.instance_group = instance_group
|
|
||||||
execution_node_msg = f' and execution node {task.execution_node}' if task.execution_node else ''
|
execution_node_msg = f' and execution node {task.execution_node}' if task.execution_node else ''
|
||||||
logger.debug(
|
logger.debug(
|
||||||
f'Submitting job {task.log_format} controlled by {task.controller_node} to instance group {instance_group.name}{execution_node_msg}.'
|
f'Submitting job {task.log_format} controlled by {task.controller_node} to instance group {instance_group.name}{execution_node_msg}.'
|
||||||
@@ -559,6 +555,7 @@ class TaskManager(TaskBase):
|
|||||||
# apply_async does a NOTIFY to the channel dispatcher is listening to
|
# apply_async does a NOTIFY to the channel dispatcher is listening to
|
||||||
# postgres will treat this as part of the transaction, which is what we want
|
# postgres will treat this as part of the transaction, which is what we want
|
||||||
if task.status != 'failed' and type(task) is not WorkflowJob:
|
if task.status != 'failed' and type(task) is not WorkflowJob:
|
||||||
|
task_actual = {'type': get_type_for_model(type(task)), 'id': task.id}
|
||||||
task_cls = task._get_task_class()
|
task_cls = task._get_task_class()
|
||||||
task_cls.apply_async(
|
task_cls.apply_async(
|
||||||
[task.pk],
|
[task.pk],
|
||||||
@@ -566,7 +563,7 @@ class TaskManager(TaskBase):
|
|||||||
queue=task.get_queue_name(),
|
queue=task.get_queue_name(),
|
||||||
uuid=task.celery_task_id,
|
uuid=task.celery_task_id,
|
||||||
callbacks=[{'task': handle_work_success.name, 'kwargs': {'task_actual': task_actual}}],
|
callbacks=[{'task': handle_work_success.name, 'kwargs': {'task_actual': task_actual}}],
|
||||||
errbacks=[{'task': handle_work_error.name, 'args': [task.celery_task_id], 'kwargs': {'subtasks': [task_actual] + dependencies}}],
|
errbacks=[{'task': handle_work_error.name, 'kwargs': {'task_actual': task_actual}}],
|
||||||
)
|
)
|
||||||
|
|
||||||
# In exception cases, like a job failing pre-start checks, we send the websocket status message
|
# In exception cases, like a job failing pre-start checks, we send the websocket status message
|
||||||
@@ -580,6 +577,7 @@ class TaskManager(TaskBase):
|
|||||||
if type(task) is WorkflowJob:
|
if type(task) is WorkflowJob:
|
||||||
ScheduleWorkflowManager().schedule()
|
ScheduleWorkflowManager().schedule()
|
||||||
self.dependency_graph.add_job(task)
|
self.dependency_graph.add_job(task)
|
||||||
|
self.tm_models.consume_capacity(task)
|
||||||
|
|
||||||
@timeit
|
@timeit
|
||||||
def process_pending_tasks(self, pending_tasks):
|
def process_pending_tasks(self, pending_tasks):
|
||||||
@@ -604,18 +602,18 @@ class TaskManager(TaskBase):
|
|||||||
if isinstance(task, WorkflowJob):
|
if isinstance(task, WorkflowJob):
|
||||||
# Previously we were tracking allow_simultaneous blocking both here and in DependencyGraph.
|
# Previously we were tracking allow_simultaneous blocking both here and in DependencyGraph.
|
||||||
# Double check that using just the DependencyGraph works for Workflows and Sliced Jobs.
|
# Double check that using just the DependencyGraph works for Workflows and Sliced Jobs.
|
||||||
self.start_task(task, None, task.get_jobs_fail_chain(), None)
|
self.start_task(task, None, None)
|
||||||
continue
|
continue
|
||||||
|
|
||||||
found_acceptable_queue = False
|
found_acceptable_queue = False
|
||||||
|
|
||||||
# Determine if there is control capacity for the task
|
# Determine if there is control capacity for the task
|
||||||
if task.capacity_type == 'control':
|
if task.capacity_type == 'control':
|
||||||
control_impact = task.task_impact + settings.AWX_CONTROL_NODE_TASK_IMPACT
|
control_impact = task.task_impact + self.control_task_impact
|
||||||
else:
|
else:
|
||||||
control_impact = settings.AWX_CONTROL_NODE_TASK_IMPACT
|
control_impact = self.control_task_impact
|
||||||
control_instance = self.instance_groups.fit_task_to_most_remaining_capacity_instance(
|
control_instance = self.tm_models.instance_groups.fit_task_to_most_remaining_capacity_instance(
|
||||||
task, instance_group_name=settings.DEFAULT_CONTROL_PLANE_QUEUE_NAME, impact=control_impact, capacity_type='control'
|
task, instance_group_name=self.controlplane_ig.name, impact=control_impact, capacity_type='control'
|
||||||
)
|
)
|
||||||
if not control_instance:
|
if not control_instance:
|
||||||
self.task_needs_capacity(task, tasks_to_update_job_explanation)
|
self.task_needs_capacity(task, tasks_to_update_job_explanation)
|
||||||
@@ -626,25 +624,29 @@ class TaskManager(TaskBase):
|
|||||||
|
|
||||||
# All task.capacity_type == 'control' jobs should run on control plane, no need to loop over instance groups
|
# All task.capacity_type == 'control' jobs should run on control plane, no need to loop over instance groups
|
||||||
if task.capacity_type == 'control':
|
if task.capacity_type == 'control':
|
||||||
|
if not self.tm_models.instance_groups[self.controlplane_ig.name].has_remaining_capacity(control_impact=True):
|
||||||
|
continue
|
||||||
task.execution_node = control_instance.hostname
|
task.execution_node = control_instance.hostname
|
||||||
execution_instance = self.instances[control_instance.hostname].obj
|
execution_instance = self.tm_models.instances[control_instance.hostname].obj
|
||||||
task.log_lifecycle("controller_node_chosen")
|
task.log_lifecycle("controller_node_chosen")
|
||||||
task.log_lifecycle("execution_node_chosen")
|
task.log_lifecycle("execution_node_chosen")
|
||||||
self.start_task(task, self.controlplane_ig, task.get_jobs_fail_chain(), execution_instance)
|
self.start_task(task, self.controlplane_ig, execution_instance)
|
||||||
found_acceptable_queue = True
|
found_acceptable_queue = True
|
||||||
continue
|
continue
|
||||||
|
|
||||||
for instance_group in self.instance_groups.get_instance_groups_from_task_cache(task):
|
for instance_group in self.tm_models.instance_groups.get_instance_groups_from_task_cache(task):
|
||||||
|
if not self.tm_models.instance_groups[instance_group.name].has_remaining_capacity(task):
|
||||||
|
continue
|
||||||
if instance_group.is_container_group:
|
if instance_group.is_container_group:
|
||||||
self.start_task(task, instance_group, task.get_jobs_fail_chain(), None)
|
self.start_task(task, instance_group, None)
|
||||||
found_acceptable_queue = True
|
found_acceptable_queue = True
|
||||||
break
|
break
|
||||||
|
|
||||||
# at this point we know the instance group is NOT a container group
|
# at this point we know the instance group is NOT a container group
|
||||||
# because if it was, it would have started the task and broke out of the loop.
|
# because if it was, it would have started the task and broke out of the loop.
|
||||||
execution_instance = self.instance_groups.fit_task_to_most_remaining_capacity_instance(
|
execution_instance = self.tm_models.instance_groups.fit_task_to_most_remaining_capacity_instance(
|
||||||
task, instance_group_name=instance_group.name, add_hybrid_control_cost=True
|
task, instance_group_name=instance_group.name, add_hybrid_control_cost=True
|
||||||
) or self.instance_groups.find_largest_idle_instance(instance_group_name=instance_group.name, capacity_type=task.capacity_type)
|
) or self.tm_models.instance_groups.find_largest_idle_instance(instance_group_name=instance_group.name, capacity_type=task.capacity_type)
|
||||||
|
|
||||||
if execution_instance:
|
if execution_instance:
|
||||||
task.execution_node = execution_instance.hostname
|
task.execution_node = execution_instance.hostname
|
||||||
@@ -660,8 +662,8 @@ class TaskManager(TaskBase):
|
|||||||
task.log_format, instance_group.name, execution_instance.hostname, execution_instance.remaining_capacity
|
task.log_format, instance_group.name, execution_instance.hostname, execution_instance.remaining_capacity
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
execution_instance = self.instances[execution_instance.hostname].obj
|
execution_instance = self.tm_models.instances[execution_instance.hostname].obj
|
||||||
self.start_task(task, instance_group, task.get_jobs_fail_chain(), execution_instance)
|
self.start_task(task, instance_group, execution_instance)
|
||||||
found_acceptable_queue = True
|
found_acceptable_queue = True
|
||||||
break
|
break
|
||||||
else:
|
else:
|
||||||
|
|||||||
@@ -15,15 +15,18 @@ logger = logging.getLogger('awx.main.scheduler')
|
|||||||
class TaskManagerInstance:
|
class TaskManagerInstance:
|
||||||
"""A class representing minimal data the task manager needs to represent an Instance."""
|
"""A class representing minimal data the task manager needs to represent an Instance."""
|
||||||
|
|
||||||
def __init__(self, obj):
|
def __init__(self, obj, **kwargs):
|
||||||
self.obj = obj
|
self.obj = obj
|
||||||
self.node_type = obj.node_type
|
self.node_type = obj.node_type
|
||||||
self.consumed_capacity = 0
|
self.consumed_capacity = 0
|
||||||
self.capacity = obj.capacity
|
self.capacity = obj.capacity
|
||||||
self.hostname = obj.hostname
|
self.hostname = obj.hostname
|
||||||
|
self.jobs_running = 0
|
||||||
|
|
||||||
def consume_capacity(self, impact):
|
def consume_capacity(self, impact, job_impact=False):
|
||||||
self.consumed_capacity += impact
|
self.consumed_capacity += impact
|
||||||
|
if job_impact:
|
||||||
|
self.jobs_running += 1
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def remaining_capacity(self):
|
def remaining_capacity(self):
|
||||||
@@ -33,9 +36,106 @@ class TaskManagerInstance:
|
|||||||
return remaining
|
return remaining
|
||||||
|
|
||||||
|
|
||||||
|
class TaskManagerInstanceGroup:
|
||||||
|
"""A class representing minimal data the task manager needs to represent an InstanceGroup."""
|
||||||
|
|
||||||
|
def __init__(self, obj, task_manager_instances=None, **kwargs):
|
||||||
|
self.name = obj.name
|
||||||
|
self.is_container_group = obj.is_container_group
|
||||||
|
self.container_group_jobs = 0
|
||||||
|
self.container_group_consumed_forks = 0
|
||||||
|
_instances = obj.instances.all()
|
||||||
|
# We want the list of TaskManagerInstance objects because these are shared across the TaskManagerInstanceGroup objects.
|
||||||
|
# This way when we consume capacity on an instance that is in multiple groups, we tabulate across all the groups correctly.
|
||||||
|
self.instances = [task_manager_instances[instance.hostname] for instance in _instances if instance.hostname in task_manager_instances]
|
||||||
|
self.instance_hostnames = tuple([instance.hostname for instance in _instances if instance.hostname in task_manager_instances])
|
||||||
|
self.max_concurrent_jobs = obj.max_concurrent_jobs
|
||||||
|
self.max_forks = obj.max_forks
|
||||||
|
self.control_task_impact = kwargs.get('control_task_impact', settings.AWX_CONTROL_NODE_TASK_IMPACT)
|
||||||
|
|
||||||
|
def consume_capacity(self, task):
|
||||||
|
"""We only consume capacity on an instance group level if it is a container group. Otherwise we consume capacity on an instance level."""
|
||||||
|
if self.is_container_group:
|
||||||
|
self.container_group_jobs += 1
|
||||||
|
self.container_group_consumed_forks += task.task_impact
|
||||||
|
else:
|
||||||
|
raise RuntimeError("We only track capacity for container groups at the instance group level. Otherwise, consume capacity on instances.")
|
||||||
|
|
||||||
|
def get_remaining_instance_capacity(self):
|
||||||
|
return sum(inst.remaining_capacity for inst in self.instances)
|
||||||
|
|
||||||
|
def get_instance_capacity(self):
|
||||||
|
return sum(inst.capacity for inst in self.instances)
|
||||||
|
|
||||||
|
def get_consumed_instance_capacity(self):
|
||||||
|
return sum(inst.consumed_capacity for inst in self.instances)
|
||||||
|
|
||||||
|
def get_instance_jobs_running(self):
|
||||||
|
return sum(inst.jobs_running for inst in self.instances)
|
||||||
|
|
||||||
|
def get_jobs_running(self):
|
||||||
|
if self.is_container_group:
|
||||||
|
return self.container_group_jobs
|
||||||
|
return sum(inst.jobs_running for inst in self.instances)
|
||||||
|
|
||||||
|
def get_capacity(self):
|
||||||
|
"""This reports any type of capacity, including that of container group jobs.
|
||||||
|
|
||||||
|
Container groups don't really have capacity, but if they have max_forks set,
|
||||||
|
we can interperet that as how much capacity the user has defined them to have.
|
||||||
|
"""
|
||||||
|
if self.is_container_group:
|
||||||
|
return self.max_forks
|
||||||
|
return self.get_instance_capacity()
|
||||||
|
|
||||||
|
def get_consumed_capacity(self):
|
||||||
|
if self.is_container_group:
|
||||||
|
return self.container_group_consumed_forks
|
||||||
|
return self.get_consumed_instance_capacity()
|
||||||
|
|
||||||
|
def get_remaining_capacity(self):
|
||||||
|
return self.get_capacity() - self.get_consumed_capacity()
|
||||||
|
|
||||||
|
def has_remaining_capacity(self, task=None, control_impact=False):
|
||||||
|
"""Pass either a task or control_impact=True to determine if the IG has capacity to run the control task or job task."""
|
||||||
|
task_impact = self.control_task_impact if control_impact else task.task_impact
|
||||||
|
job_impact = 0 if control_impact else 1
|
||||||
|
task_string = f"task {task.log_format} with impact of {task_impact}" if task else f"control task with impact of {task_impact}"
|
||||||
|
|
||||||
|
# We only want to loop over instances if self.max_concurrent_jobs is set
|
||||||
|
if self.max_concurrent_jobs == 0:
|
||||||
|
# Override the calculated remaining capacity, because when max_concurrent_jobs == 0 we don't enforce any max
|
||||||
|
remaining_jobs = 0
|
||||||
|
else:
|
||||||
|
remaining_jobs = self.max_concurrent_jobs - self.get_jobs_running() - job_impact
|
||||||
|
|
||||||
|
# We only want to loop over instances if self.max_forks is set
|
||||||
|
if self.max_forks == 0:
|
||||||
|
# Override the calculated remaining capacity, because when max_forks == 0 we don't enforce any max
|
||||||
|
remaining_forks = 0
|
||||||
|
else:
|
||||||
|
remaining_forks = self.max_forks - self.get_consumed_capacity() - task_impact
|
||||||
|
|
||||||
|
if remaining_jobs < 0 or remaining_forks < 0:
|
||||||
|
# A value less than zero means the task will not fit on the group
|
||||||
|
if remaining_jobs < 0:
|
||||||
|
logger.debug(f"{task_string} cannot fit on instance group {self.name} with {remaining_jobs} remaining jobs")
|
||||||
|
if remaining_forks < 0:
|
||||||
|
logger.debug(f"{task_string} cannot fit on instance group {self.name} with {remaining_forks} remaining forks")
|
||||||
|
return False
|
||||||
|
|
||||||
|
# Returning true means there is enough remaining capacity on the group to run the task (or no instance group level limits are being set)
|
||||||
|
logger.debug(f"{task_string} can fit on instance group {self.name} with {remaining_forks} remaining forks and {remaining_jobs}")
|
||||||
|
return True
|
||||||
|
|
||||||
|
|
||||||
class TaskManagerInstances:
|
class TaskManagerInstances:
|
||||||
def __init__(self, active_tasks, instances=None, instance_fields=('node_type', 'capacity', 'hostname', 'enabled')):
|
def __init__(self, instances=None, instance_fields=('node_type', 'capacity', 'hostname', 'enabled'), **kwargs):
|
||||||
self.instances_by_hostname = dict()
|
self.instances_by_hostname = dict()
|
||||||
|
self.instance_groups_container_group_jobs = dict()
|
||||||
|
self.instance_groups_container_group_consumed_forks = dict()
|
||||||
|
self.control_task_impact = kwargs.get('control_task_impact', settings.AWX_CONTROL_NODE_TASK_IMPACT)
|
||||||
|
|
||||||
if instances is None:
|
if instances is None:
|
||||||
instances = (
|
instances = (
|
||||||
Instance.objects.filter(hostname__isnull=False, node_state=Instance.States.READY, enabled=True)
|
Instance.objects.filter(hostname__isnull=False, node_state=Instance.States.READY, enabled=True)
|
||||||
@@ -43,18 +143,15 @@ class TaskManagerInstances:
|
|||||||
.only('node_type', 'node_state', 'capacity', 'hostname', 'enabled')
|
.only('node_type', 'node_state', 'capacity', 'hostname', 'enabled')
|
||||||
)
|
)
|
||||||
for instance in instances:
|
for instance in instances:
|
||||||
self.instances_by_hostname[instance.hostname] = TaskManagerInstance(instance)
|
self.instances_by_hostname[instance.hostname] = TaskManagerInstance(instance, **kwargs)
|
||||||
|
|
||||||
# initialize remaining capacity based on currently waiting and running tasks
|
def consume_capacity(self, task):
|
||||||
for task in active_tasks:
|
|
||||||
if task.status not in ['waiting', 'running']:
|
|
||||||
continue
|
|
||||||
control_instance = self.instances_by_hostname.get(task.controller_node, '')
|
control_instance = self.instances_by_hostname.get(task.controller_node, '')
|
||||||
execution_instance = self.instances_by_hostname.get(task.execution_node, '')
|
execution_instance = self.instances_by_hostname.get(task.execution_node, '')
|
||||||
if execution_instance and execution_instance.node_type in ('hybrid', 'execution'):
|
if execution_instance and execution_instance.node_type in ('hybrid', 'execution'):
|
||||||
self.instances_by_hostname[task.execution_node].consume_capacity(task.task_impact)
|
self.instances_by_hostname[task.execution_node].consume_capacity(task.task_impact, job_impact=True)
|
||||||
if control_instance and control_instance.node_type in ('hybrid', 'control'):
|
if control_instance and control_instance.node_type in ('hybrid', 'control'):
|
||||||
self.instances_by_hostname[task.controller_node].consume_capacity(settings.AWX_CONTROL_NODE_TASK_IMPACT)
|
self.instances_by_hostname[task.controller_node].consume_capacity(self.control_task_impact)
|
||||||
|
|
||||||
def __getitem__(self, hostname):
|
def __getitem__(self, hostname):
|
||||||
return self.instances_by_hostname.get(hostname)
|
return self.instances_by_hostname.get(hostname)
|
||||||
@@ -64,42 +161,57 @@ class TaskManagerInstances:
|
|||||||
|
|
||||||
|
|
||||||
class TaskManagerInstanceGroups:
|
class TaskManagerInstanceGroups:
|
||||||
"""A class representing minimal data the task manager needs to represent an InstanceGroup."""
|
"""A class representing minimal data the task manager needs to represent all the InstanceGroups."""
|
||||||
|
|
||||||
def __init__(self, instances_by_hostname=None, instance_groups=None, instance_groups_queryset=None):
|
def __init__(self, task_manager_instances=None, instance_groups=None, instance_groups_queryset=None, **kwargs):
|
||||||
self.instance_groups = dict()
|
self.instance_groups = dict()
|
||||||
|
self.task_manager_instances = task_manager_instances if task_manager_instances is not None else TaskManagerInstances()
|
||||||
self.controlplane_ig = None
|
self.controlplane_ig = None
|
||||||
self.pk_ig_map = dict()
|
self.pk_ig_map = dict()
|
||||||
|
self.control_task_impact = kwargs.get('control_task_impact', settings.AWX_CONTROL_NODE_TASK_IMPACT)
|
||||||
|
self.controlplane_ig_name = kwargs.get('controlplane_ig_name', settings.DEFAULT_CONTROL_PLANE_QUEUE_NAME)
|
||||||
|
|
||||||
if instance_groups is not None: # for testing
|
if instance_groups is not None: # for testing
|
||||||
self.instance_groups = instance_groups
|
self.instance_groups = {ig.name: TaskManagerInstanceGroup(ig, self.task_manager_instances, **kwargs) for ig in instance_groups}
|
||||||
|
self.pk_ig_map = {ig.pk: ig for ig in instance_groups}
|
||||||
else:
|
else:
|
||||||
if instance_groups_queryset is None:
|
if instance_groups_queryset is None:
|
||||||
instance_groups_queryset = InstanceGroup.objects.prefetch_related('instances').only('name', 'instances')
|
instance_groups_queryset = InstanceGroup.objects.prefetch_related('instances').only(
|
||||||
for instance_group in instance_groups_queryset:
|
'name', 'instances', 'max_concurrent_jobs', 'max_forks', 'is_container_group'
|
||||||
if instance_group.name == settings.DEFAULT_CONTROL_PLANE_QUEUE_NAME:
|
|
||||||
self.controlplane_ig = instance_group
|
|
||||||
self.instance_groups[instance_group.name] = dict(
|
|
||||||
instances=[
|
|
||||||
instances_by_hostname[instance.hostname] for instance in instance_group.instances.all() if instance.hostname in instances_by_hostname
|
|
||||||
],
|
|
||||||
)
|
)
|
||||||
|
for instance_group in instance_groups_queryset:
|
||||||
|
if instance_group.name == self.controlplane_ig_name:
|
||||||
|
self.controlplane_ig = instance_group
|
||||||
|
self.instance_groups[instance_group.name] = TaskManagerInstanceGroup(instance_group, self.task_manager_instances, **kwargs)
|
||||||
self.pk_ig_map[instance_group.pk] = instance_group
|
self.pk_ig_map[instance_group.pk] = instance_group
|
||||||
|
|
||||||
|
def __getitem__(self, ig_name):
|
||||||
|
return self.instance_groups.get(ig_name)
|
||||||
|
|
||||||
|
def __contains__(self, ig_name):
|
||||||
|
return ig_name in self.instance_groups
|
||||||
|
|
||||||
def get_remaining_capacity(self, group_name):
|
def get_remaining_capacity(self, group_name):
|
||||||
instances = self.instance_groups[group_name]['instances']
|
return self.instance_groups[group_name].get_remaining_instance_capacity()
|
||||||
return sum(inst.remaining_capacity for inst in instances)
|
|
||||||
|
|
||||||
def get_consumed_capacity(self, group_name):
|
def get_consumed_capacity(self, group_name):
|
||||||
instances = self.instance_groups[group_name]['instances']
|
return self.instance_groups[group_name].get_consumed_capacity()
|
||||||
return sum(inst.consumed_capacity for inst in instances)
|
|
||||||
|
def get_jobs_running(self, group_name):
|
||||||
|
return self.instance_groups[group_name].get_jobs_running()
|
||||||
|
|
||||||
|
def get_capacity(self, group_name):
|
||||||
|
return self.instance_groups[group_name].get_capacity()
|
||||||
|
|
||||||
|
def get_instances(self, group_name):
|
||||||
|
return self.instance_groups[group_name].instances
|
||||||
|
|
||||||
def fit_task_to_most_remaining_capacity_instance(self, task, instance_group_name, impact=None, capacity_type=None, add_hybrid_control_cost=False):
|
def fit_task_to_most_remaining_capacity_instance(self, task, instance_group_name, impact=None, capacity_type=None, add_hybrid_control_cost=False):
|
||||||
impact = impact if impact else task.task_impact
|
impact = impact if impact else task.task_impact
|
||||||
capacity_type = capacity_type if capacity_type else task.capacity_type
|
capacity_type = capacity_type if capacity_type else task.capacity_type
|
||||||
instance_most_capacity = None
|
instance_most_capacity = None
|
||||||
most_remaining_capacity = -1
|
most_remaining_capacity = -1
|
||||||
instances = self.instance_groups[instance_group_name]['instances']
|
instances = self.instance_groups[instance_group_name].instances
|
||||||
|
|
||||||
for i in instances:
|
for i in instances:
|
||||||
if i.node_type not in (capacity_type, 'hybrid'):
|
if i.node_type not in (capacity_type, 'hybrid'):
|
||||||
@@ -107,7 +219,7 @@ class TaskManagerInstanceGroups:
|
|||||||
would_be_remaining = i.remaining_capacity - impact
|
would_be_remaining = i.remaining_capacity - impact
|
||||||
# hybrid nodes _always_ control their own tasks
|
# hybrid nodes _always_ control their own tasks
|
||||||
if add_hybrid_control_cost and i.node_type == 'hybrid':
|
if add_hybrid_control_cost and i.node_type == 'hybrid':
|
||||||
would_be_remaining -= settings.AWX_CONTROL_NODE_TASK_IMPACT
|
would_be_remaining -= self.control_task_impact
|
||||||
if would_be_remaining >= 0 and (instance_most_capacity is None or would_be_remaining > most_remaining_capacity):
|
if would_be_remaining >= 0 and (instance_most_capacity is None or would_be_remaining > most_remaining_capacity):
|
||||||
instance_most_capacity = i
|
instance_most_capacity = i
|
||||||
most_remaining_capacity = would_be_remaining
|
most_remaining_capacity = would_be_remaining
|
||||||
@@ -115,10 +227,13 @@ class TaskManagerInstanceGroups:
|
|||||||
|
|
||||||
def find_largest_idle_instance(self, instance_group_name, capacity_type='execution'):
|
def find_largest_idle_instance(self, instance_group_name, capacity_type='execution'):
|
||||||
largest_instance = None
|
largest_instance = None
|
||||||
instances = self.instance_groups[instance_group_name]['instances']
|
instances = self.instance_groups[instance_group_name].instances
|
||||||
for i in instances:
|
for i in instances:
|
||||||
if i.node_type not in (capacity_type, 'hybrid'):
|
if i.node_type not in (capacity_type, 'hybrid'):
|
||||||
continue
|
continue
|
||||||
|
if i.capacity <= 0:
|
||||||
|
# We don't want to select an idle instance with 0 capacity
|
||||||
|
continue
|
||||||
if (hasattr(i, 'jobs_running') and i.jobs_running == 0) or i.remaining_capacity == i.capacity:
|
if (hasattr(i, 'jobs_running') and i.jobs_running == 0) or i.remaining_capacity == i.capacity:
|
||||||
if largest_instance is None:
|
if largest_instance is None:
|
||||||
largest_instance = i
|
largest_instance = i
|
||||||
@@ -139,3 +254,56 @@ class TaskManagerInstanceGroups:
|
|||||||
logger.warn(f"No instance groups in cache exist, defaulting to global instance groups for task {task}")
|
logger.warn(f"No instance groups in cache exist, defaulting to global instance groups for task {task}")
|
||||||
return task.global_instance_groups
|
return task.global_instance_groups
|
||||||
return igs
|
return igs
|
||||||
|
|
||||||
|
|
||||||
|
class TaskManagerModels:
|
||||||
|
def __init__(self, **kwargs):
|
||||||
|
# We want to avoid calls to settings over and over in loops, so cache this information here
|
||||||
|
kwargs['control_task_impact'] = kwargs.get('control_task_impact', settings.AWX_CONTROL_NODE_TASK_IMPACT)
|
||||||
|
kwargs['controlplane_ig_name'] = kwargs.get('controlplane_ig_name', settings.DEFAULT_CONTROL_PLANE_QUEUE_NAME)
|
||||||
|
self.instances = TaskManagerInstances(**kwargs)
|
||||||
|
self.instance_groups = TaskManagerInstanceGroups(task_manager_instances=self.instances, **kwargs)
|
||||||
|
|
||||||
|
@classmethod
|
||||||
|
def init_with_consumed_capacity(cls, **kwargs):
|
||||||
|
tmm = cls(**kwargs)
|
||||||
|
tasks = kwargs.get('tasks', None)
|
||||||
|
|
||||||
|
if tasks is None:
|
||||||
|
instance_group_queryset = kwargs.get('instance_groups_queryset', None)
|
||||||
|
# No tasks were provided, so we will fetch them from the database
|
||||||
|
task_status_filter_list = kwargs.get('task_status_filter_list', ['running', 'waiting'])
|
||||||
|
task_fields = kwargs.get('task_fields', ('task_impact', 'controller_node', 'execution_node', 'instance_group'))
|
||||||
|
from awx.main.models import UnifiedJob
|
||||||
|
|
||||||
|
if instance_group_queryset is not None:
|
||||||
|
logger.debug("******************INSTANCE GROUP QUERYSET PASSED -- FILTERING TASKS ****************************")
|
||||||
|
# Sometimes things like the serializer pass a queryset that looks at not all instance groups. in this case,
|
||||||
|
# we also need to filter the tasks we look at
|
||||||
|
tasks = UnifiedJob.objects.filter(status__in=task_status_filter_list, instance_group__in=[ig.id for ig in instance_group_queryset]).only(
|
||||||
|
*task_fields
|
||||||
|
)
|
||||||
|
else:
|
||||||
|
# No instance group query set, look at all tasks in whole system
|
||||||
|
tasks = UnifiedJob.objects.filter(status__in=task_status_filter_list).only(*task_fields)
|
||||||
|
|
||||||
|
for task in tasks:
|
||||||
|
tmm.consume_capacity(task)
|
||||||
|
|
||||||
|
return tmm
|
||||||
|
|
||||||
|
def consume_capacity(self, task):
|
||||||
|
# Consume capacity on instances, which bubbles up to instance groups they are a member of
|
||||||
|
self.instances.consume_capacity(task)
|
||||||
|
|
||||||
|
# For container group jobs, additionally we must account for capacity consumed since
|
||||||
|
# The container groups have no instances to look at to track how many jobs/forks are consumed
|
||||||
|
if task.instance_group_id:
|
||||||
|
if not task.instance_group_id in self.instance_groups.pk_ig_map.keys():
|
||||||
|
logger.warn(
|
||||||
|
f"Task {task.log_format} assigned {task.instance_group_id} but this instance group not present in map of instance groups{self.instance_groups.pk_ig_map.keys()}"
|
||||||
|
)
|
||||||
|
else:
|
||||||
|
ig = self.instance_groups.pk_ig_map[task.instance_group_id]
|
||||||
|
if ig.is_container_group:
|
||||||
|
self.instance_groups[ig.name].consume_capacity(task)
|
||||||
|
|||||||
@@ -63,7 +63,7 @@ def read_receptor_config():
|
|||||||
|
|
||||||
def work_signing_enabled(config_data):
|
def work_signing_enabled(config_data):
|
||||||
for section in config_data:
|
for section in config_data:
|
||||||
if 'work-verification' in section:
|
if 'work-signing' in section:
|
||||||
return True
|
return True
|
||||||
return False
|
return False
|
||||||
|
|
||||||
@@ -411,9 +411,11 @@ class AWXReceptorJob:
|
|||||||
unit_status = receptor_ctl.simple_command(f'work status {self.unit_id}')
|
unit_status = receptor_ctl.simple_command(f'work status {self.unit_id}')
|
||||||
detail = unit_status.get('Detail', None)
|
detail = unit_status.get('Detail', None)
|
||||||
state_name = unit_status.get('StateName', None)
|
state_name = unit_status.get('StateName', None)
|
||||||
|
stdout_size = unit_status.get('StdoutSize', 0)
|
||||||
except Exception:
|
except Exception:
|
||||||
detail = ''
|
detail = ''
|
||||||
state_name = ''
|
state_name = ''
|
||||||
|
stdout_size = 0
|
||||||
logger.exception(f'An error was encountered while getting status for work unit {self.unit_id}')
|
logger.exception(f'An error was encountered while getting status for work unit {self.unit_id}')
|
||||||
|
|
||||||
if 'exceeded quota' in detail:
|
if 'exceeded quota' in detail:
|
||||||
@@ -424,8 +426,15 @@ class AWXReceptorJob:
|
|||||||
return
|
return
|
||||||
|
|
||||||
try:
|
try:
|
||||||
resultsock = receptor_ctl.get_work_results(self.unit_id, return_sockfile=True)
|
receptor_output = ''
|
||||||
lines = resultsock.readlines()
|
if state_name == 'Failed' and self.task.runner_callback.event_ct == 0:
|
||||||
|
# if receptor work unit failed and no events were emitted, work results may
|
||||||
|
# contain useful information about why the job failed. In case stdout is
|
||||||
|
# massive, only ask for last 1000 bytes
|
||||||
|
startpos = max(stdout_size - 1000, 0)
|
||||||
|
resultsock, resultfile = receptor_ctl.get_work_results(self.unit_id, startpos=startpos, return_socket=True, return_sockfile=True)
|
||||||
|
resultsock.setblocking(False) # this makes resultfile reads non blocking
|
||||||
|
lines = resultfile.readlines()
|
||||||
receptor_output = b"".join(lines).decode()
|
receptor_output = b"".join(lines).decode()
|
||||||
if receptor_output:
|
if receptor_output:
|
||||||
self.task.runner_callback.delay_update(result_traceback=receptor_output)
|
self.task.runner_callback.delay_update(result_traceback=receptor_output)
|
||||||
|
|||||||
@@ -52,6 +52,7 @@ from awx.main.constants import ACTIVE_STATES
|
|||||||
from awx.main.dispatch.publish import task
|
from awx.main.dispatch.publish import task
|
||||||
from awx.main.dispatch import get_local_queuename, reaper
|
from awx.main.dispatch import get_local_queuename, reaper
|
||||||
from awx.main.utils.common import (
|
from awx.main.utils.common import (
|
||||||
|
get_type_for_model,
|
||||||
ignore_inventory_computed_fields,
|
ignore_inventory_computed_fields,
|
||||||
ignore_inventory_group_removal,
|
ignore_inventory_group_removal,
|
||||||
ScheduleWorkflowManager,
|
ScheduleWorkflowManager,
|
||||||
@@ -720,45 +721,43 @@ def handle_work_success(task_actual):
|
|||||||
|
|
||||||
|
|
||||||
@task(queue=get_local_queuename)
|
@task(queue=get_local_queuename)
|
||||||
def handle_work_error(task_id, *args, **kwargs):
|
def handle_work_error(task_actual):
|
||||||
subtasks = kwargs.get('subtasks', None)
|
|
||||||
logger.debug('Executing error task id %s, subtasks: %s' % (task_id, str(subtasks)))
|
|
||||||
first_instance = None
|
|
||||||
first_instance_type = ''
|
|
||||||
if subtasks is not None:
|
|
||||||
for each_task in subtasks:
|
|
||||||
try:
|
try:
|
||||||
instance = UnifiedJob.get_instance_by_type(each_task['type'], each_task['id'])
|
instance = UnifiedJob.get_instance_by_type(task_actual['type'], task_actual['id'])
|
||||||
if not instance:
|
|
||||||
# Unknown task type
|
|
||||||
logger.warning("Unknown task type: {}".format(each_task['type']))
|
|
||||||
continue
|
|
||||||
except ObjectDoesNotExist:
|
except ObjectDoesNotExist:
|
||||||
logger.warning('Missing {} `{}` in error callback.'.format(each_task['type'], each_task['id']))
|
logger.warning('Missing {} `{}` in error callback.'.format(task_actual['type'], task_actual['id']))
|
||||||
continue
|
return
|
||||||
|
if not instance:
|
||||||
|
return
|
||||||
|
|
||||||
if first_instance is None:
|
subtasks = instance.get_jobs_fail_chain() # reverse of dependent_jobs mostly
|
||||||
first_instance = instance
|
logger.debug(f'Executing error task id {task_actual["id"]}, subtasks: {[subtask.id for subtask in subtasks]}')
|
||||||
first_instance_type = each_task['type']
|
|
||||||
|
|
||||||
if instance.celery_task_id != task_id and not instance.cancel_flag and not instance.status in ('successful', 'failed'):
|
deps_of_deps = {}
|
||||||
instance.status = 'failed'
|
|
||||||
instance.failed = True
|
for subtask in subtasks:
|
||||||
if not instance.job_explanation:
|
if subtask.celery_task_id != instance.celery_task_id and not subtask.cancel_flag and not subtask.status in ('successful', 'failed'):
|
||||||
instance.job_explanation = 'Previous Task Failed: {"job_type": "%s", "job_name": "%s", "job_id": "%s"}' % (
|
# If there are multiple in the dependency chain, A->B->C, and this was called for A, blame B for clarity
|
||||||
first_instance_type,
|
blame_job = deps_of_deps.get(subtask.id, instance)
|
||||||
first_instance.name,
|
subtask.status = 'failed'
|
||||||
first_instance.id,
|
subtask.failed = True
|
||||||
|
if not subtask.job_explanation:
|
||||||
|
subtask.job_explanation = 'Previous Task Failed: {"job_type": "%s", "job_name": "%s", "job_id": "%s"}' % (
|
||||||
|
get_type_for_model(type(blame_job)),
|
||||||
|
blame_job.name,
|
||||||
|
blame_job.id,
|
||||||
)
|
)
|
||||||
instance.save()
|
subtask.save()
|
||||||
instance.websocket_emit_status("failed")
|
subtask.websocket_emit_status("failed")
|
||||||
|
|
||||||
|
for sub_subtask in subtask.get_jobs_fail_chain():
|
||||||
|
deps_of_deps[sub_subtask.id] = subtask
|
||||||
|
|
||||||
# We only send 1 job complete message since all the job completion message
|
# We only send 1 job complete message since all the job completion message
|
||||||
# handling does is trigger the scheduler. If we extend the functionality of
|
# handling does is trigger the scheduler. If we extend the functionality of
|
||||||
# what the job complete message handler does then we may want to send a
|
# what the job complete message handler does then we may want to send a
|
||||||
# completion event for each job here.
|
# completion event for each job here.
|
||||||
if first_instance:
|
schedule_manager_success_or_error(instance)
|
||||||
schedule_manager_success_or_error(first_instance)
|
|
||||||
|
|
||||||
|
|
||||||
@task(queue=get_local_queuename)
|
@task(queue=get_local_queuename)
|
||||||
|
|||||||
@@ -3,5 +3,6 @@
|
|||||||
"ANSIBLE_TRANSFORM_INVALID_GROUP_CHARS": "never",
|
"ANSIBLE_TRANSFORM_INVALID_GROUP_CHARS": "never",
|
||||||
"AWS_ACCESS_KEY_ID": "fooo",
|
"AWS_ACCESS_KEY_ID": "fooo",
|
||||||
"AWS_SECRET_ACCESS_KEY": "fooo",
|
"AWS_SECRET_ACCESS_KEY": "fooo",
|
||||||
"AWS_SECURITY_TOKEN": "fooo"
|
"AWS_SECURITY_TOKEN": "fooo",
|
||||||
|
"AWS_SESSION_TOKEN": "fooo"
|
||||||
}
|
}
|
||||||
@@ -171,13 +171,17 @@ class TestKeyRegeneration:
|
|||||||
|
|
||||||
def test_use_custom_key_with_empty_tower_secret_key_env_var(self):
|
def test_use_custom_key_with_empty_tower_secret_key_env_var(self):
|
||||||
os.environ['TOWER_SECRET_KEY'] = ''
|
os.environ['TOWER_SECRET_KEY'] = ''
|
||||||
new_key = call_command('regenerate_secret_key', '--use-custom-key')
|
with pytest.raises(SystemExit) as e:
|
||||||
assert settings.SECRET_KEY != new_key
|
call_command('regenerate_secret_key', '--use-custom-key')
|
||||||
|
assert e.type == SystemExit
|
||||||
|
assert e.value.code == 1
|
||||||
|
|
||||||
def test_use_custom_key_with_no_tower_secret_key_env_var(self):
|
def test_use_custom_key_with_no_tower_secret_key_env_var(self):
|
||||||
os.environ.pop('TOWER_SECRET_KEY', None)
|
os.environ.pop('TOWER_SECRET_KEY', None)
|
||||||
new_key = call_command('regenerate_secret_key', '--use-custom-key')
|
with pytest.raises(SystemExit) as e:
|
||||||
assert settings.SECRET_KEY != new_key
|
call_command('regenerate_secret_key', '--use-custom-key')
|
||||||
|
assert e.type == SystemExit
|
||||||
|
assert e.value.code == 1
|
||||||
|
|
||||||
def test_with_tower_secret_key_env_var(self):
|
def test_with_tower_secret_key_env_var(self):
|
||||||
custom_key = 'MXSq9uqcwezBOChl/UfmbW1k4op+bC+FQtwPqgJ1u9XV'
|
custom_key = 'MXSq9uqcwezBOChl/UfmbW1k4op+bC+FQtwPqgJ1u9XV'
|
||||||
|
|||||||
@@ -4,7 +4,7 @@ from awx.main.models import (
|
|||||||
Instance,
|
Instance,
|
||||||
InstanceGroup,
|
InstanceGroup,
|
||||||
)
|
)
|
||||||
from awx.main.scheduler.task_manager_models import TaskManagerInstanceGroups, TaskManagerInstances
|
from awx.main.scheduler.task_manager_models import TaskManagerInstanceGroups
|
||||||
|
|
||||||
|
|
||||||
class TestInstanceGroupInstanceMapping(TransactionTestCase):
|
class TestInstanceGroupInstanceMapping(TransactionTestCase):
|
||||||
@@ -23,11 +23,10 @@ class TestInstanceGroupInstanceMapping(TransactionTestCase):
|
|||||||
def test_mapping(self):
|
def test_mapping(self):
|
||||||
self.sample_cluster()
|
self.sample_cluster()
|
||||||
with self.assertNumQueries(3):
|
with self.assertNumQueries(3):
|
||||||
instances = TaskManagerInstances([]) # empty task list
|
instance_groups = TaskManagerInstanceGroups()
|
||||||
instance_groups = TaskManagerInstanceGroups(instances_by_hostname=instances)
|
|
||||||
|
|
||||||
ig_instance_map = instance_groups.instance_groups
|
ig_instance_map = instance_groups.instance_groups
|
||||||
|
|
||||||
assert set(i.hostname for i in ig_instance_map['ig_small']['instances']) == set(['i1'])
|
assert set(i.hostname for i in ig_instance_map['ig_small'].instances) == set(['i1'])
|
||||||
assert set(i.hostname for i in ig_instance_map['ig_large']['instances']) == set(['i2', 'i3'])
|
assert set(i.hostname for i in ig_instance_map['ig_large'].instances) == set(['i2', 'i3'])
|
||||||
assert set(i.hostname for i in ig_instance_map['default']['instances']) == set(['i2'])
|
assert set(i.hostname for i in ig_instance_map['default'].instances) == set(['i2'])
|
||||||
|
|||||||
@@ -10,6 +10,10 @@ from awx.main.utils import (
|
|||||||
create_temporary_fifo,
|
create_temporary_fifo,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
from awx.main.scheduler import TaskManager
|
||||||
|
|
||||||
|
from . import create_job
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture
|
@pytest.fixture
|
||||||
def containerized_job(default_instance_group, kube_credential, job_template_factory):
|
def containerized_job(default_instance_group, kube_credential, job_template_factory):
|
||||||
@@ -34,6 +38,50 @@ def test_containerized_job(containerized_job):
|
|||||||
assert containerized_job.instance_group.credential.kubernetes
|
assert containerized_job.instance_group.credential.kubernetes
|
||||||
|
|
||||||
|
|
||||||
|
@pytest.mark.django_db
|
||||||
|
def test_max_concurrent_jobs_blocks_start_of_new_jobs(controlplane_instance_group, containerized_job, mocker):
|
||||||
|
"""Construct a scenario where only 1 job will fit within the max_concurrent_jobs of the container group.
|
||||||
|
|
||||||
|
Since max_concurrent_jobs is set to 1, even though 2 jobs are in pending
|
||||||
|
and would be launched into the container group, only one will be started.
|
||||||
|
"""
|
||||||
|
containerized_job.unified_job_template.allow_simultaneous = True
|
||||||
|
containerized_job.unified_job_template.save()
|
||||||
|
default_instance_group = containerized_job.instance_group
|
||||||
|
default_instance_group.max_concurrent_jobs = 1
|
||||||
|
default_instance_group.save()
|
||||||
|
task_impact = 1
|
||||||
|
# Create a second job that should not be scheduled at first, blocked by the other
|
||||||
|
create_job(containerized_job.unified_job_template)
|
||||||
|
tm = TaskManager()
|
||||||
|
with mock.patch('awx.main.models.Job.task_impact', new_callable=mock.PropertyMock) as mock_task_impact:
|
||||||
|
mock_task_impact.return_value = task_impact
|
||||||
|
with mock.patch.object(TaskManager, "start_task", wraps=tm.start_task) as mock_job:
|
||||||
|
tm.schedule()
|
||||||
|
mock_job.assert_called_once()
|
||||||
|
|
||||||
|
|
||||||
|
@pytest.mark.django_db
|
||||||
|
def test_max_forks_blocks_start_of_new_jobs(controlplane_instance_group, containerized_job, mocker):
|
||||||
|
"""Construct a scenario where only 1 job will fit within the max_forks of the container group.
|
||||||
|
|
||||||
|
In this case, we set the container_group max_forks to 10, and make the task_impact of a job 6.
|
||||||
|
Therefore, only 1 job will fit within the max of 10.
|
||||||
|
"""
|
||||||
|
containerized_job.unified_job_template.allow_simultaneous = True
|
||||||
|
containerized_job.unified_job_template.save()
|
||||||
|
default_instance_group = containerized_job.instance_group
|
||||||
|
default_instance_group.max_forks = 10
|
||||||
|
# Create a second job that should not be scheduled
|
||||||
|
create_job(containerized_job.unified_job_template)
|
||||||
|
tm = TaskManager()
|
||||||
|
with mock.patch('awx.main.models.Job.task_impact', new_callable=mock.PropertyMock) as mock_task_impact:
|
||||||
|
mock_task_impact.return_value = 6
|
||||||
|
with mock.patch("awx.main.scheduler.TaskManager.start_task"):
|
||||||
|
tm.schedule()
|
||||||
|
tm.start_task.assert_called_once()
|
||||||
|
|
||||||
|
|
||||||
@pytest.mark.django_db
|
@pytest.mark.django_db
|
||||||
def test_kubectl_ssl_verification(containerized_job, default_job_execution_environment):
|
def test_kubectl_ssl_verification(containerized_job, default_job_execution_environment):
|
||||||
containerized_job.execution_environment = default_job_execution_environment
|
containerized_job.execution_environment = default_job_execution_environment
|
||||||
|
|||||||
@@ -23,7 +23,7 @@ def test_multi_group_basic_job_launch(instance_factory, controlplane_instance_gr
|
|||||||
mock_task_impact.return_value = 500
|
mock_task_impact.return_value = 500
|
||||||
with mocker.patch("awx.main.scheduler.TaskManager.start_task"):
|
with mocker.patch("awx.main.scheduler.TaskManager.start_task"):
|
||||||
TaskManager().schedule()
|
TaskManager().schedule()
|
||||||
TaskManager.start_task.assert_has_calls([mock.call(j1, ig1, [], i1), mock.call(j2, ig2, [], i2)])
|
TaskManager.start_task.assert_has_calls([mock.call(j1, ig1, i1), mock.call(j2, ig2, i2)])
|
||||||
|
|
||||||
|
|
||||||
@pytest.mark.django_db
|
@pytest.mark.django_db
|
||||||
@@ -54,7 +54,7 @@ def test_multi_group_with_shared_dependency(instance_factory, controlplane_insta
|
|||||||
DependencyManager().schedule()
|
DependencyManager().schedule()
|
||||||
TaskManager().schedule()
|
TaskManager().schedule()
|
||||||
pu = p.project_updates.first()
|
pu = p.project_updates.first()
|
||||||
TaskManager.start_task.assert_called_once_with(pu, controlplane_instance_group, [j1, j2], controlplane_instance_group.instances.all()[0])
|
TaskManager.start_task.assert_called_once_with(pu, controlplane_instance_group, controlplane_instance_group.instances.all()[0])
|
||||||
pu.finished = pu.created + timedelta(seconds=1)
|
pu.finished = pu.created + timedelta(seconds=1)
|
||||||
pu.status = "successful"
|
pu.status = "successful"
|
||||||
pu.save()
|
pu.save()
|
||||||
@@ -62,8 +62,8 @@ def test_multi_group_with_shared_dependency(instance_factory, controlplane_insta
|
|||||||
DependencyManager().schedule()
|
DependencyManager().schedule()
|
||||||
TaskManager().schedule()
|
TaskManager().schedule()
|
||||||
|
|
||||||
TaskManager.start_task.assert_any_call(j1, ig1, [], i1)
|
TaskManager.start_task.assert_any_call(j1, ig1, i1)
|
||||||
TaskManager.start_task.assert_any_call(j2, ig2, [], i2)
|
TaskManager.start_task.assert_any_call(j2, ig2, i2)
|
||||||
assert TaskManager.start_task.call_count == 2
|
assert TaskManager.start_task.call_count == 2
|
||||||
|
|
||||||
|
|
||||||
@@ -75,7 +75,7 @@ def test_workflow_job_no_instancegroup(workflow_job_template_factory, controlpla
|
|||||||
wfj.save()
|
wfj.save()
|
||||||
with mocker.patch("awx.main.scheduler.TaskManager.start_task"):
|
with mocker.patch("awx.main.scheduler.TaskManager.start_task"):
|
||||||
TaskManager().schedule()
|
TaskManager().schedule()
|
||||||
TaskManager.start_task.assert_called_once_with(wfj, None, [], None)
|
TaskManager.start_task.assert_called_once_with(wfj, None, None)
|
||||||
assert wfj.instance_group is None
|
assert wfj.instance_group is None
|
||||||
|
|
||||||
|
|
||||||
@@ -150,7 +150,7 @@ def test_failover_group_run(instance_factory, controlplane_instance_group, mocke
|
|||||||
mock_task_impact.return_value = 500
|
mock_task_impact.return_value = 500
|
||||||
with mock.patch.object(TaskManager, "start_task", wraps=tm.start_task) as mock_job:
|
with mock.patch.object(TaskManager, "start_task", wraps=tm.start_task) as mock_job:
|
||||||
tm.schedule()
|
tm.schedule()
|
||||||
mock_job.assert_has_calls([mock.call(j1, ig1, [], i1), mock.call(j1_1, ig2, [], i2)])
|
mock_job.assert_has_calls([mock.call(j1, ig1, i1), mock.call(j1_1, ig2, i2)])
|
||||||
assert mock_job.call_count == 2
|
assert mock_job.call_count == 2
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
@@ -18,7 +18,7 @@ def test_single_job_scheduler_launch(hybrid_instance, controlplane_instance_grou
|
|||||||
j = create_job(objects.job_template)
|
j = create_job(objects.job_template)
|
||||||
with mocker.patch("awx.main.scheduler.TaskManager.start_task"):
|
with mocker.patch("awx.main.scheduler.TaskManager.start_task"):
|
||||||
TaskManager().schedule()
|
TaskManager().schedule()
|
||||||
TaskManager.start_task.assert_called_once_with(j, controlplane_instance_group, [], instance)
|
TaskManager.start_task.assert_called_once_with(j, controlplane_instance_group, instance)
|
||||||
|
|
||||||
|
|
||||||
@pytest.mark.django_db
|
@pytest.mark.django_db
|
||||||
@@ -240,12 +240,82 @@ def test_multi_jt_capacity_blocking(hybrid_instance, job_template_factory, mocke
|
|||||||
mock_task_impact.return_value = 505
|
mock_task_impact.return_value = 505
|
||||||
with mock.patch.object(TaskManager, "start_task", wraps=tm.start_task) as mock_job:
|
with mock.patch.object(TaskManager, "start_task", wraps=tm.start_task) as mock_job:
|
||||||
tm.schedule()
|
tm.schedule()
|
||||||
mock_job.assert_called_once_with(j1, controlplane_instance_group, [], instance)
|
mock_job.assert_called_once_with(j1, controlplane_instance_group, instance)
|
||||||
j1.status = "successful"
|
j1.status = "successful"
|
||||||
j1.save()
|
j1.save()
|
||||||
with mock.patch.object(TaskManager, "start_task", wraps=tm.start_task) as mock_job:
|
with mock.patch.object(TaskManager, "start_task", wraps=tm.start_task) as mock_job:
|
||||||
tm.schedule()
|
tm.schedule()
|
||||||
mock_job.assert_called_once_with(j2, controlplane_instance_group, [], instance)
|
mock_job.assert_called_once_with(j2, controlplane_instance_group, instance)
|
||||||
|
|
||||||
|
|
||||||
|
@pytest.mark.django_db
|
||||||
|
def test_max_concurrent_jobs_ig_capacity_blocking(hybrid_instance, job_template_factory, mocker):
|
||||||
|
"""When max_concurrent_jobs of an instance group is more restrictive than capacity of instances, enforce max_concurrent_jobs."""
|
||||||
|
instance = hybrid_instance
|
||||||
|
controlplane_instance_group = instance.rampart_groups.first()
|
||||||
|
# We will expect only 1 job to be started
|
||||||
|
controlplane_instance_group.max_concurrent_jobs = 1
|
||||||
|
controlplane_instance_group.save()
|
||||||
|
num_jobs = 3
|
||||||
|
jobs = []
|
||||||
|
for i in range(num_jobs):
|
||||||
|
jobs.append(
|
||||||
|
create_job(job_template_factory(f'jt{i}', organization=f'org{i}', project=f'proj{i}', inventory=f'inv{i}', credential=f'cred{i}').job_template)
|
||||||
|
)
|
||||||
|
tm = TaskManager()
|
||||||
|
task_impact = 1
|
||||||
|
|
||||||
|
# Sanity check that multiple jobs would run if not for the max_concurrent_jobs setting.
|
||||||
|
assert task_impact * num_jobs < controlplane_instance_group.capacity
|
||||||
|
tm = TaskManager()
|
||||||
|
with mock.patch('awx.main.models.Job.task_impact', new_callable=mock.PropertyMock) as mock_task_impact:
|
||||||
|
mock_task_impact.return_value = task_impact
|
||||||
|
with mock.patch.object(TaskManager, "start_task", wraps=tm.start_task) as mock_job:
|
||||||
|
tm.schedule()
|
||||||
|
mock_job.assert_called_once()
|
||||||
|
jobs[0].status = 'running'
|
||||||
|
jobs[0].controller_node = instance.hostname
|
||||||
|
jobs[0].execution_node = instance.hostname
|
||||||
|
jobs[0].instance_group = controlplane_instance_group
|
||||||
|
jobs[0].save()
|
||||||
|
|
||||||
|
# while that job is running, we should not start another job
|
||||||
|
with mock.patch('awx.main.models.Job.task_impact', new_callable=mock.PropertyMock) as mock_task_impact:
|
||||||
|
mock_task_impact.return_value = task_impact
|
||||||
|
with mock.patch.object(TaskManager, "start_task", wraps=tm.start_task) as mock_job:
|
||||||
|
tm.schedule()
|
||||||
|
mock_job.assert_not_called()
|
||||||
|
# now job is done, we should start one of the two other jobs
|
||||||
|
jobs[0].status = 'successful'
|
||||||
|
jobs[0].save()
|
||||||
|
with mock.patch('awx.main.models.Job.task_impact', new_callable=mock.PropertyMock) as mock_task_impact:
|
||||||
|
mock_task_impact.return_value = task_impact
|
||||||
|
with mock.patch.object(TaskManager, "start_task", wraps=tm.start_task) as mock_job:
|
||||||
|
tm.schedule()
|
||||||
|
mock_job.assert_called_once()
|
||||||
|
|
||||||
|
|
||||||
|
@pytest.mark.django_db
|
||||||
|
def test_max_forks_ig_capacity_blocking(hybrid_instance, job_template_factory, mocker):
|
||||||
|
"""When max_forks of an instance group is less than the capacity of instances, enforce max_forks."""
|
||||||
|
instance = hybrid_instance
|
||||||
|
controlplane_instance_group = instance.rampart_groups.first()
|
||||||
|
controlplane_instance_group.max_forks = 15
|
||||||
|
controlplane_instance_group.save()
|
||||||
|
task_impact = 10
|
||||||
|
num_jobs = 2
|
||||||
|
# Sanity check that 2 jobs would run if not for the max_forks setting.
|
||||||
|
assert controlplane_instance_group.max_forks < controlplane_instance_group.capacity
|
||||||
|
assert task_impact * num_jobs > controlplane_instance_group.max_forks
|
||||||
|
assert task_impact * num_jobs < controlplane_instance_group.capacity
|
||||||
|
for i in range(num_jobs):
|
||||||
|
create_job(job_template_factory(f'jt{i}', organization=f'org{i}', project=f'proj{i}', inventory=f'inv{i}', credential=f'cred{i}').job_template)
|
||||||
|
tm = TaskManager()
|
||||||
|
with mock.patch('awx.main.models.Job.task_impact', new_callable=mock.PropertyMock) as mock_task_impact:
|
||||||
|
mock_task_impact.return_value = task_impact
|
||||||
|
with mock.patch.object(TaskManager, "start_task", wraps=tm.start_task) as mock_job:
|
||||||
|
tm.schedule()
|
||||||
|
mock_job.assert_called_once()
|
||||||
|
|
||||||
|
|
||||||
@pytest.mark.django_db
|
@pytest.mark.django_db
|
||||||
@@ -267,12 +337,12 @@ def test_single_job_dependencies_project_launch(controlplane_instance_group, job
|
|||||||
pu = [x for x in p.project_updates.all()]
|
pu = [x for x in p.project_updates.all()]
|
||||||
assert len(pu) == 1
|
assert len(pu) == 1
|
||||||
TaskManager().schedule()
|
TaskManager().schedule()
|
||||||
TaskManager.start_task.assert_called_once_with(pu[0], controlplane_instance_group, [j], instance)
|
TaskManager.start_task.assert_called_once_with(pu[0], controlplane_instance_group, instance)
|
||||||
pu[0].status = "successful"
|
pu[0].status = "successful"
|
||||||
pu[0].save()
|
pu[0].save()
|
||||||
with mock.patch("awx.main.scheduler.TaskManager.start_task"):
|
with mock.patch("awx.main.scheduler.TaskManager.start_task"):
|
||||||
TaskManager().schedule()
|
TaskManager().schedule()
|
||||||
TaskManager.start_task.assert_called_once_with(j, controlplane_instance_group, [], instance)
|
TaskManager.start_task.assert_called_once_with(j, controlplane_instance_group, instance)
|
||||||
|
|
||||||
|
|
||||||
@pytest.mark.django_db
|
@pytest.mark.django_db
|
||||||
@@ -295,12 +365,12 @@ def test_single_job_dependencies_inventory_update_launch(controlplane_instance_g
|
|||||||
iu = [x for x in ii.inventory_updates.all()]
|
iu = [x for x in ii.inventory_updates.all()]
|
||||||
assert len(iu) == 1
|
assert len(iu) == 1
|
||||||
TaskManager().schedule()
|
TaskManager().schedule()
|
||||||
TaskManager.start_task.assert_called_once_with(iu[0], controlplane_instance_group, [j], instance)
|
TaskManager.start_task.assert_called_once_with(iu[0], controlplane_instance_group, instance)
|
||||||
iu[0].status = "successful"
|
iu[0].status = "successful"
|
||||||
iu[0].save()
|
iu[0].save()
|
||||||
with mock.patch("awx.main.scheduler.TaskManager.start_task"):
|
with mock.patch("awx.main.scheduler.TaskManager.start_task"):
|
||||||
TaskManager().schedule()
|
TaskManager().schedule()
|
||||||
TaskManager.start_task.assert_called_once_with(j, controlplane_instance_group, [], instance)
|
TaskManager.start_task.assert_called_once_with(j, controlplane_instance_group, instance)
|
||||||
|
|
||||||
|
|
||||||
@pytest.mark.django_db
|
@pytest.mark.django_db
|
||||||
@@ -342,7 +412,7 @@ def test_job_dependency_with_already_updated(controlplane_instance_group, job_te
|
|||||||
mock_iu.assert_not_called()
|
mock_iu.assert_not_called()
|
||||||
with mock.patch("awx.main.scheduler.TaskManager.start_task"):
|
with mock.patch("awx.main.scheduler.TaskManager.start_task"):
|
||||||
TaskManager().schedule()
|
TaskManager().schedule()
|
||||||
TaskManager.start_task.assert_called_once_with(j, controlplane_instance_group, [], instance)
|
TaskManager.start_task.assert_called_once_with(j, controlplane_instance_group, instance)
|
||||||
|
|
||||||
|
|
||||||
@pytest.mark.django_db
|
@pytest.mark.django_db
|
||||||
@@ -372,9 +442,7 @@ def test_shared_dependencies_launch(controlplane_instance_group, job_template_fa
|
|||||||
TaskManager().schedule()
|
TaskManager().schedule()
|
||||||
pu = p.project_updates.first()
|
pu = p.project_updates.first()
|
||||||
iu = ii.inventory_updates.first()
|
iu = ii.inventory_updates.first()
|
||||||
TaskManager.start_task.assert_has_calls(
|
TaskManager.start_task.assert_has_calls([mock.call(iu, controlplane_instance_group, instance), mock.call(pu, controlplane_instance_group, instance)])
|
||||||
[mock.call(iu, controlplane_instance_group, [j1, j2], instance), mock.call(pu, controlplane_instance_group, [j1, j2], instance)]
|
|
||||||
)
|
|
||||||
pu.status = "successful"
|
pu.status = "successful"
|
||||||
pu.finished = pu.created + timedelta(seconds=1)
|
pu.finished = pu.created + timedelta(seconds=1)
|
||||||
pu.save()
|
pu.save()
|
||||||
@@ -383,9 +451,7 @@ def test_shared_dependencies_launch(controlplane_instance_group, job_template_fa
|
|||||||
iu.save()
|
iu.save()
|
||||||
with mock.patch("awx.main.scheduler.TaskManager.start_task"):
|
with mock.patch("awx.main.scheduler.TaskManager.start_task"):
|
||||||
TaskManager().schedule()
|
TaskManager().schedule()
|
||||||
TaskManager.start_task.assert_has_calls(
|
TaskManager.start_task.assert_has_calls([mock.call(j1, controlplane_instance_group, instance), mock.call(j2, controlplane_instance_group, instance)])
|
||||||
[mock.call(j1, controlplane_instance_group, [], instance), mock.call(j2, controlplane_instance_group, [], instance)]
|
|
||||||
)
|
|
||||||
pu = [x for x in p.project_updates.all()]
|
pu = [x for x in p.project_updates.all()]
|
||||||
iu = [x for x in ii.inventory_updates.all()]
|
iu = [x for x in ii.inventory_updates.all()]
|
||||||
assert len(pu) == 1
|
assert len(pu) == 1
|
||||||
@@ -409,7 +475,7 @@ def test_job_not_blocking_project_update(controlplane_instance_group, job_templa
|
|||||||
project_update.status = "pending"
|
project_update.status = "pending"
|
||||||
project_update.save()
|
project_update.save()
|
||||||
TaskManager().schedule()
|
TaskManager().schedule()
|
||||||
TaskManager.start_task.assert_called_once_with(project_update, controlplane_instance_group, [], instance)
|
TaskManager.start_task.assert_called_once_with(project_update, controlplane_instance_group, instance)
|
||||||
|
|
||||||
|
|
||||||
@pytest.mark.django_db
|
@pytest.mark.django_db
|
||||||
@@ -433,7 +499,7 @@ def test_job_not_blocking_inventory_update(controlplane_instance_group, job_temp
|
|||||||
|
|
||||||
DependencyManager().schedule()
|
DependencyManager().schedule()
|
||||||
TaskManager().schedule()
|
TaskManager().schedule()
|
||||||
TaskManager.start_task.assert_called_once_with(inventory_update, controlplane_instance_group, [], instance)
|
TaskManager.start_task.assert_called_once_with(inventory_update, controlplane_instance_group, instance)
|
||||||
|
|
||||||
|
|
||||||
@pytest.mark.django_db
|
@pytest.mark.django_db
|
||||||
|
|||||||
@@ -5,8 +5,8 @@ import tempfile
|
|||||||
import shutil
|
import shutil
|
||||||
|
|
||||||
from awx.main.tasks.jobs import RunJob
|
from awx.main.tasks.jobs import RunJob
|
||||||
from awx.main.tasks.system import execution_node_health_check, _cleanup_images_and_files
|
from awx.main.tasks.system import execution_node_health_check, _cleanup_images_and_files, handle_work_error
|
||||||
from awx.main.models import Instance, Job
|
from awx.main.models import Instance, Job, InventoryUpdate, ProjectUpdate
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture
|
@pytest.fixture
|
||||||
@@ -74,3 +74,17 @@ def test_does_not_run_reaped_job(mocker, mock_me):
|
|||||||
job.refresh_from_db()
|
job.refresh_from_db()
|
||||||
assert job.status == 'failed'
|
assert job.status == 'failed'
|
||||||
mock_run.assert_not_called()
|
mock_run.assert_not_called()
|
||||||
|
|
||||||
|
|
||||||
|
@pytest.mark.django_db
|
||||||
|
def test_handle_work_error_nested(project, inventory_source):
|
||||||
|
pu = ProjectUpdate.objects.create(status='failed', project=project, celery_task_id='1234')
|
||||||
|
iu = InventoryUpdate.objects.create(status='pending', inventory_source=inventory_source, source='scm')
|
||||||
|
job = Job.objects.create(status='pending')
|
||||||
|
iu.dependent_jobs.add(pu)
|
||||||
|
job.dependent_jobs.add(pu, iu)
|
||||||
|
handle_work_error({'type': 'project_update', 'id': pu.id})
|
||||||
|
iu.refresh_from_db()
|
||||||
|
job.refresh_from_db()
|
||||||
|
assert iu.job_explanation == f'Previous Task Failed: {{"job_type": "project_update", "job_name": "", "job_id": "{pu.id}"}}'
|
||||||
|
assert job.job_explanation == f'Previous Task Failed: {{"job_type": "inventory_update", "job_name": "", "job_id": "{iu.id}"}}'
|
||||||
|
|||||||
@@ -1,10 +1,7 @@
|
|||||||
import pytest
|
import pytest
|
||||||
from unittest import mock
|
|
||||||
from unittest.mock import Mock
|
|
||||||
from decimal import Decimal
|
from decimal import Decimal
|
||||||
|
|
||||||
from awx.main.models import InstanceGroup, Instance
|
from awx.main.models import Instance
|
||||||
from awx.main.scheduler.task_manager_models import TaskManagerInstanceGroups
|
|
||||||
|
|
||||||
|
|
||||||
@pytest.mark.parametrize('capacity_adjustment', [0.0, 0.25, 0.5, 0.75, 1, 1.5, 3])
|
@pytest.mark.parametrize('capacity_adjustment', [0.0, 0.25, 0.5, 0.75, 1, 1.5, 3])
|
||||||
@@ -17,83 +14,6 @@ def test_capacity_adjustment_no_save(capacity_adjustment):
|
|||||||
assert inst.capacity == (float(inst.capacity_adjustment) * abs(inst.mem_capacity - inst.cpu_capacity) + min(inst.mem_capacity, inst.cpu_capacity))
|
assert inst.capacity == (float(inst.capacity_adjustment) * abs(inst.mem_capacity - inst.cpu_capacity) + min(inst.mem_capacity, inst.cpu_capacity))
|
||||||
|
|
||||||
|
|
||||||
def T(impact):
|
|
||||||
j = mock.Mock(spec_set=['task_impact', 'capacity_type'])
|
|
||||||
j.task_impact = impact
|
|
||||||
j.capacity_type = 'execution'
|
|
||||||
return j
|
|
||||||
|
|
||||||
|
|
||||||
def Is(param):
|
|
||||||
"""
|
|
||||||
param:
|
|
||||||
[remaining_capacity1, remaining_capacity2, remaining_capacity3, ...]
|
|
||||||
[(jobs_running1, capacity1), (jobs_running2, capacity2), (jobs_running3, capacity3), ...]
|
|
||||||
"""
|
|
||||||
|
|
||||||
instances = []
|
|
||||||
if isinstance(param[0], tuple):
|
|
||||||
for (jobs_running, capacity) in param:
|
|
||||||
inst = Mock()
|
|
||||||
inst.capacity = capacity
|
|
||||||
inst.jobs_running = jobs_running
|
|
||||||
inst.node_type = 'execution'
|
|
||||||
instances.append(inst)
|
|
||||||
else:
|
|
||||||
for i in param:
|
|
||||||
inst = Mock()
|
|
||||||
inst.remaining_capacity = i
|
|
||||||
inst.node_type = 'execution'
|
|
||||||
instances.append(inst)
|
|
||||||
return instances
|
|
||||||
|
|
||||||
|
|
||||||
class TestInstanceGroup(object):
|
|
||||||
@pytest.mark.parametrize(
|
|
||||||
'task,instances,instance_fit_index,reason',
|
|
||||||
[
|
|
||||||
(T(100), Is([100]), 0, "Only one, pick it"),
|
|
||||||
(T(100), Is([100, 100]), 0, "Two equally good fits, pick the first"),
|
|
||||||
(T(100), Is([50, 100]), 1, "First instance not as good as second instance"),
|
|
||||||
(T(100), Is([50, 0, 20, 100, 100, 100, 30, 20]), 3, "Pick Instance [3] as it is the first that the task fits in."),
|
|
||||||
(T(100), Is([50, 0, 20, 99, 11, 1, 5, 99]), None, "The task don't a fit, you must a quit!"),
|
|
||||||
],
|
|
||||||
)
|
|
||||||
def test_fit_task_to_most_remaining_capacity_instance(self, task, instances, instance_fit_index, reason):
|
|
||||||
InstanceGroup(id=10)
|
|
||||||
tm_igs = TaskManagerInstanceGroups(instance_groups={'controlplane': {'instances': instances}})
|
|
||||||
|
|
||||||
instance_picked = tm_igs.fit_task_to_most_remaining_capacity_instance(task, 'controlplane')
|
|
||||||
|
|
||||||
if instance_fit_index is None:
|
|
||||||
assert instance_picked is None, reason
|
|
||||||
else:
|
|
||||||
assert instance_picked == instances[instance_fit_index], reason
|
|
||||||
|
|
||||||
@pytest.mark.parametrize(
|
|
||||||
'instances,instance_fit_index,reason',
|
|
||||||
[
|
|
||||||
(Is([(0, 100)]), 0, "One idle instance, pick it"),
|
|
||||||
(Is([(1, 100)]), None, "One un-idle instance, pick nothing"),
|
|
||||||
(Is([(0, 100), (0, 200), (1, 500), (0, 700)]), 3, "Pick the largest idle instance"),
|
|
||||||
(Is([(0, 100), (0, 200), (1, 10000), (0, 700), (0, 699)]), 3, "Pick the largest idle instance"),
|
|
||||||
(Is([(0, 0)]), None, "One idle but down instance, don't pick it"),
|
|
||||||
],
|
|
||||||
)
|
|
||||||
def test_find_largest_idle_instance(self, instances, instance_fit_index, reason):
|
|
||||||
def filter_offline_instances(*args):
|
|
||||||
return filter(lambda i: i.capacity > 0, instances)
|
|
||||||
|
|
||||||
InstanceGroup(id=10)
|
|
||||||
instances_online_only = filter_offline_instances(instances)
|
|
||||||
tm_igs = TaskManagerInstanceGroups(instance_groups={'controlplane': {'instances': instances_online_only}})
|
|
||||||
|
|
||||||
if instance_fit_index is None:
|
|
||||||
assert tm_igs.find_largest_idle_instance('controlplane') is None, reason
|
|
||||||
else:
|
|
||||||
assert tm_igs.find_largest_idle_instance('controlplane') == instances[instance_fit_index], reason
|
|
||||||
|
|
||||||
|
|
||||||
def test_cleanup_params_defaults():
|
def test_cleanup_params_defaults():
|
||||||
inst = Instance(hostname='foobar')
|
inst = Instance(hostname='foobar')
|
||||||
assert inst.get_cleanup_task_kwargs(exclude_strings=['awx_423_']) == {'exclude_strings': ['awx_423_'], 'file_pattern': '/tmp/awx_*_*', 'grace_period': 60}
|
assert inst.get_cleanup_task_kwargs(exclude_strings=['awx_423_']) == {'exclude_strings': ['awx_423_'], 'file_pattern': '/tmp/awx_*_*', 'grace_period': 60}
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
import pytest
|
import pytest
|
||||||
|
|
||||||
from awx.main.scheduler.task_manager_models import TaskManagerInstanceGroups, TaskManagerInstances
|
from awx.main.scheduler.task_manager_models import TaskManagerModels
|
||||||
|
|
||||||
|
|
||||||
class FakeMeta(object):
|
class FakeMeta(object):
|
||||||
@@ -16,18 +16,19 @@ class FakeObject(object):
|
|||||||
|
|
||||||
|
|
||||||
class Job(FakeObject):
|
class Job(FakeObject):
|
||||||
task_impact = 43
|
def __init__(self, **kwargs):
|
||||||
is_container_group_task = False
|
self.task_impact = kwargs.get('task_impact', 43)
|
||||||
controller_node = ''
|
self.is_container_group_task = kwargs.get('is_container_group_task', False)
|
||||||
execution_node = ''
|
self.controller_node = kwargs.get('controller_node', '')
|
||||||
|
self.execution_node = kwargs.get('execution_node', '')
|
||||||
|
self.instance_group = kwargs.get('instance_group', None)
|
||||||
|
self.instance_group_id = self.instance_group.id if self.instance_group else None
|
||||||
|
self.capacity_type = kwargs.get('capacity_type', 'execution')
|
||||||
|
|
||||||
def log_format(self):
|
def log_format(self):
|
||||||
return 'job 382 (fake)'
|
return 'job 382 (fake)'
|
||||||
|
|
||||||
|
|
||||||
@pytest.fixture
|
|
||||||
def sample_cluster():
|
|
||||||
def stand_up_cluster():
|
|
||||||
class Instances(FakeObject):
|
class Instances(FakeObject):
|
||||||
def add(self, *args):
|
def add(self, *args):
|
||||||
for instance in args:
|
for instance in args:
|
||||||
@@ -36,18 +37,43 @@ def sample_cluster():
|
|||||||
def all(self):
|
def all(self):
|
||||||
return self.obj.instance_list
|
return self.obj.instance_list
|
||||||
|
|
||||||
|
|
||||||
class InstanceGroup(FakeObject):
|
class InstanceGroup(FakeObject):
|
||||||
def __init__(self, **kwargs):
|
def __init__(self, **kwargs):
|
||||||
super(InstanceGroup, self).__init__(**kwargs)
|
super(InstanceGroup, self).__init__(**kwargs)
|
||||||
self.instance_list = []
|
self.instance_list = []
|
||||||
|
self.pk = self.id = kwargs.get('id', 1)
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def instances(self):
|
def instances(self):
|
||||||
mgr = Instances(obj=self)
|
mgr = Instances(obj=self)
|
||||||
return mgr
|
return mgr
|
||||||
|
|
||||||
|
@property
|
||||||
|
def is_container_group(self):
|
||||||
|
return False
|
||||||
|
|
||||||
|
@property
|
||||||
|
def max_concurrent_jobs(self):
|
||||||
|
return 0
|
||||||
|
|
||||||
|
@property
|
||||||
|
def max_forks(self):
|
||||||
|
return 0
|
||||||
|
|
||||||
|
|
||||||
class Instance(FakeObject):
|
class Instance(FakeObject):
|
||||||
pass
|
def __init__(self, **kwargs):
|
||||||
|
self.node_type = kwargs.get('node_type', 'hybrid')
|
||||||
|
self.capacity = kwargs.get('capacity', 0)
|
||||||
|
self.hostname = kwargs.get('hostname', 'fakehostname')
|
||||||
|
self.consumed_capacity = 0
|
||||||
|
self.jobs_running = 0
|
||||||
|
|
||||||
|
|
||||||
|
@pytest.fixture
|
||||||
|
def sample_cluster():
|
||||||
|
def stand_up_cluster():
|
||||||
|
|
||||||
ig_small = InstanceGroup(name='ig_small')
|
ig_small = InstanceGroup(name='ig_small')
|
||||||
ig_large = InstanceGroup(name='ig_large')
|
ig_large = InstanceGroup(name='ig_large')
|
||||||
@@ -66,14 +92,12 @@ def sample_cluster():
|
|||||||
@pytest.fixture
|
@pytest.fixture
|
||||||
def create_ig_manager():
|
def create_ig_manager():
|
||||||
def _rf(ig_list, tasks):
|
def _rf(ig_list, tasks):
|
||||||
instances = TaskManagerInstances(tasks, instances=set(inst for ig in ig_list for inst in ig.instance_list))
|
tm_models = TaskManagerModels.init_with_consumed_capacity(
|
||||||
|
tasks=tasks,
|
||||||
seed_igs = {}
|
instances=set(inst for ig in ig_list for inst in ig.instance_list),
|
||||||
for ig in ig_list:
|
instance_groups=ig_list,
|
||||||
seed_igs[ig.name] = {'instances': [instances[inst.hostname] for inst in ig.instance_list]}
|
)
|
||||||
|
return tm_models.instance_groups
|
||||||
instance_groups = TaskManagerInstanceGroups(instance_groups=seed_igs)
|
|
||||||
return instance_groups
|
|
||||||
|
|
||||||
return _rf
|
return _rf
|
||||||
|
|
||||||
@@ -126,3 +150,75 @@ def test_RBAC_reduced_filter(sample_cluster, create_ig_manager):
|
|||||||
# Cross-links between groups not visible to current user,
|
# Cross-links between groups not visible to current user,
|
||||||
# so a naieve accounting of capacities is returned instead
|
# so a naieve accounting of capacities is returned instead
|
||||||
assert instance_groups_mgr.get_consumed_capacity('default') == 43
|
assert instance_groups_mgr.get_consumed_capacity('default') == 43
|
||||||
|
|
||||||
|
|
||||||
|
def Is(param):
|
||||||
|
"""
|
||||||
|
param:
|
||||||
|
[remaining_capacity1, remaining_capacity2, remaining_capacity3, ...]
|
||||||
|
[(jobs_running1, capacity1), (jobs_running2, capacity2), (jobs_running3, capacity3), ...]
|
||||||
|
"""
|
||||||
|
|
||||||
|
instances = []
|
||||||
|
if isinstance(param[0], tuple):
|
||||||
|
for index, (jobs_running, capacity) in enumerate(param):
|
||||||
|
inst = Instance(capacity=capacity, node_type='execution', hostname=f'fakehost-{index}')
|
||||||
|
inst.jobs_running = jobs_running
|
||||||
|
instances.append(inst)
|
||||||
|
else:
|
||||||
|
for index, capacity in enumerate(param):
|
||||||
|
inst = Instance(capacity=capacity, node_type='execution', hostname=f'fakehost-{index}')
|
||||||
|
inst.node_type = 'execution'
|
||||||
|
instances.append(inst)
|
||||||
|
return instances
|
||||||
|
|
||||||
|
|
||||||
|
class TestSelectBestInstanceForTask(object):
|
||||||
|
@pytest.mark.parametrize(
|
||||||
|
'task,instances,instance_fit_index,reason',
|
||||||
|
[
|
||||||
|
(Job(task_impact=100), Is([100]), 0, "Only one, pick it"),
|
||||||
|
(Job(task_impact=100), Is([100, 100]), 0, "Two equally good fits, pick the first"),
|
||||||
|
(Job(task_impact=100), Is([50, 100]), 1, "First instance not as good as second instance"),
|
||||||
|
(Job(task_impact=100), Is([50, 0, 20, 100, 100, 100, 30, 20]), 3, "Pick Instance [3] as it is the first that the task fits in."),
|
||||||
|
(Job(task_impact=100), Is([50, 0, 20, 99, 11, 1, 5, 99]), None, "The task don't a fit, you must a quit!"),
|
||||||
|
],
|
||||||
|
)
|
||||||
|
def test_fit_task_to_most_remaining_capacity_instance(self, task, instances, instance_fit_index, reason):
|
||||||
|
ig = InstanceGroup(id=10, name='controlplane')
|
||||||
|
tasks = []
|
||||||
|
for instance in instances:
|
||||||
|
ig.instances.add(instance)
|
||||||
|
for _ in range(instance.jobs_running):
|
||||||
|
tasks.append(Job(execution_node=instance.hostname, controller_node=instance.hostname, instance_group=ig))
|
||||||
|
tm_models = TaskManagerModels.init_with_consumed_capacity(tasks=tasks, instances=instances, instance_groups=[ig])
|
||||||
|
instance_picked = tm_models.instance_groups.fit_task_to_most_remaining_capacity_instance(task, 'controlplane')
|
||||||
|
|
||||||
|
if instance_fit_index is None:
|
||||||
|
assert instance_picked is None, reason
|
||||||
|
else:
|
||||||
|
assert instance_picked.hostname == instances[instance_fit_index].hostname, reason
|
||||||
|
|
||||||
|
@pytest.mark.parametrize(
|
||||||
|
'instances,instance_fit_index,reason',
|
||||||
|
[
|
||||||
|
(Is([(0, 100)]), 0, "One idle instance, pick it"),
|
||||||
|
(Is([(1, 100)]), None, "One un-idle instance, pick nothing"),
|
||||||
|
(Is([(0, 100), (0, 200), (1, 500), (0, 700)]), 3, "Pick the largest idle instance"),
|
||||||
|
(Is([(0, 100), (0, 200), (1, 10000), (0, 700), (0, 699)]), 3, "Pick the largest idle instance"),
|
||||||
|
(Is([(0, 0)]), None, "One idle but down instance, don't pick it"),
|
||||||
|
],
|
||||||
|
)
|
||||||
|
def test_find_largest_idle_instance(self, instances, instance_fit_index, reason):
|
||||||
|
ig = InstanceGroup(id=10, name='controlplane')
|
||||||
|
tasks = []
|
||||||
|
for instance in instances:
|
||||||
|
ig.instances.add(instance)
|
||||||
|
for _ in range(instance.jobs_running):
|
||||||
|
tasks.append(Job(execution_node=instance.hostname, controller_node=instance.hostname, instance_group=ig))
|
||||||
|
tm_models = TaskManagerModels.init_with_consumed_capacity(tasks=tasks, instances=instances, instance_groups=[ig])
|
||||||
|
|
||||||
|
if instance_fit_index is None:
|
||||||
|
assert tm_models.instance_groups.find_largest_idle_instance('controlplane') is None, reason
|
||||||
|
else:
|
||||||
|
assert tm_models.instance_groups.find_largest_idle_instance('controlplane').hostname == instances[instance_fit_index].hostname, reason
|
||||||
|
|||||||
@@ -118,7 +118,7 @@ class WebsocketTask:
|
|||||||
logger.warning(f"Connection from {self.name} to {self.remote_host} timed out.")
|
logger.warning(f"Connection from {self.name} to {self.remote_host} timed out.")
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
# Early on, this is our canary. I'm not sure what exceptions we can really encounter.
|
# Early on, this is our canary. I'm not sure what exceptions we can really encounter.
|
||||||
logger.warning(f"Connection from {self.name} to {self.remote_host} failed for unknown reason: '{e}'.")
|
logger.exception(f"Connection from {self.name} to {self.remote_host} failed for unknown reason: '{e}'.")
|
||||||
else:
|
else:
|
||||||
logger.warning(f"Connection from {self.name} to {self.remote_host} list.")
|
logger.warning(f"Connection from {self.name} to {self.remote_host} list.")
|
||||||
|
|
||||||
|
|||||||
@@ -304,11 +304,13 @@ INSTALLED_APPS = [
|
|||||||
'django.contrib.messages',
|
'django.contrib.messages',
|
||||||
'django.contrib.sessions',
|
'django.contrib.sessions',
|
||||||
'django.contrib.sites',
|
'django.contrib.sites',
|
||||||
|
# daphne has to be installed before django.contrib.staticfiles for the app to startup
|
||||||
|
# According to channels 4.0 docs you install daphne instead of channels now
|
||||||
|
'daphne',
|
||||||
'django.contrib.staticfiles',
|
'django.contrib.staticfiles',
|
||||||
'oauth2_provider',
|
'oauth2_provider',
|
||||||
'rest_framework',
|
'rest_framework',
|
||||||
'django_extensions',
|
'django_extensions',
|
||||||
'channels',
|
|
||||||
'polymorphic',
|
'polymorphic',
|
||||||
'taggit',
|
'taggit',
|
||||||
'social_django',
|
'social_django',
|
||||||
@@ -851,6 +853,7 @@ LOGGING = {
|
|||||||
'awx.main.signals': {'level': 'INFO'}, # very verbose debug-level logs
|
'awx.main.signals': {'level': 'INFO'}, # very verbose debug-level logs
|
||||||
'awx.api.permissions': {'level': 'INFO'}, # very verbose debug-level logs
|
'awx.api.permissions': {'level': 'INFO'}, # very verbose debug-level logs
|
||||||
'awx.analytics': {'handlers': ['external_logger'], 'level': 'INFO', 'propagate': False},
|
'awx.analytics': {'handlers': ['external_logger'], 'level': 'INFO', 'propagate': False},
|
||||||
|
'awx.analytics.broadcast_websocket': {'handlers': ['console', 'file', 'wsbroadcast', 'external_logger'], 'level': 'INFO', 'propagate': False},
|
||||||
'awx.analytics.performance': {'handlers': ['console', 'file', 'tower_warnings', 'external_logger'], 'level': 'DEBUG', 'propagate': False},
|
'awx.analytics.performance': {'handlers': ['console', 'file', 'tower_warnings', 'external_logger'], 'level': 'DEBUG', 'propagate': False},
|
||||||
'awx.analytics.job_lifecycle': {'handlers': ['console', 'job_lifecycle'], 'level': 'DEBUG', 'propagate': False},
|
'awx.analytics.job_lifecycle': {'handlers': ['console', 'job_lifecycle'], 'level': 'DEBUG', 'propagate': False},
|
||||||
'django_auth_ldap': {'handlers': ['console', 'file', 'tower_warnings'], 'level': 'DEBUG'},
|
'django_auth_ldap': {'handlers': ['console', 'file', 'tower_warnings'], 'level': 'DEBUG'},
|
||||||
@@ -983,6 +986,13 @@ DJANGO_GUID = {'GUID_HEADER_NAME': 'X-API-Request-Id'}
|
|||||||
DEFAULT_EXECUTION_QUEUE_NAME = 'default'
|
DEFAULT_EXECUTION_QUEUE_NAME = 'default'
|
||||||
# pod spec used when the default execution queue is a container group, e.g. when deploying on k8s/ocp with the operator
|
# pod spec used when the default execution queue is a container group, e.g. when deploying on k8s/ocp with the operator
|
||||||
DEFAULT_EXECUTION_QUEUE_POD_SPEC_OVERRIDE = ''
|
DEFAULT_EXECUTION_QUEUE_POD_SPEC_OVERRIDE = ''
|
||||||
|
# Max number of concurrently consumed forks for the default execution queue
|
||||||
|
# Zero means no limit
|
||||||
|
DEFAULT_EXECUTION_QUEUE_MAX_FORKS = 0
|
||||||
|
# Max number of concurrently running jobs for the default execution queue
|
||||||
|
# Zero means no limit
|
||||||
|
DEFAULT_EXECUTION_QUEUE_MAX_CONCURRENT_JOBS = 0
|
||||||
|
|
||||||
# Name of the default controlplane queue
|
# Name of the default controlplane queue
|
||||||
DEFAULT_CONTROL_PLANE_QUEUE_NAME = 'controlplane'
|
DEFAULT_CONTROL_PLANE_QUEUE_NAME = 'controlplane'
|
||||||
|
|
||||||
|
|||||||
@@ -114,7 +114,7 @@ if 'sqlite3' not in DATABASES['default']['ENGINE']: # noqa
|
|||||||
# this needs to stay at the bottom of this file
|
# this needs to stay at the bottom of this file
|
||||||
try:
|
try:
|
||||||
if os.getenv('AWX_KUBE_DEVEL', False):
|
if os.getenv('AWX_KUBE_DEVEL', False):
|
||||||
include(optional('minikube.py'), scope=locals())
|
include(optional('development_kube.py'), scope=locals())
|
||||||
else:
|
else:
|
||||||
include(optional('local_*.py'), scope=locals())
|
include(optional('local_*.py'), scope=locals())
|
||||||
except ImportError:
|
except ImportError:
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
BROADCAST_WEBSOCKET_SECRET = '🤖starscream🤖'
|
BROADCAST_WEBSOCKET_SECRET = '🤖starscream🤖'
|
||||||
BROADCAST_WEBSOCKET_PORT = 8013
|
BROADCAST_WEBSOCKET_PORT = 8052
|
||||||
BROADCAST_WEBSOCKET_VERIFY_CERT = False
|
BROADCAST_WEBSOCKET_VERIFY_CERT = False
|
||||||
BROADCAST_WEBSOCKET_PROTOCOL = 'http'
|
BROADCAST_WEBSOCKET_PROTOCOL = 'http'
|
||||||
79
awx/ui/package-lock.json
generated
79
awx/ui/package-lock.json
generated
@@ -8,7 +8,7 @@
|
|||||||
"dependencies": {
|
"dependencies": {
|
||||||
"@lingui/react": "3.14.0",
|
"@lingui/react": "3.14.0",
|
||||||
"@patternfly/patternfly": "4.217.1",
|
"@patternfly/patternfly": "4.217.1",
|
||||||
"@patternfly/react-core": "^4.250.1",
|
"@patternfly/react-core": "^4.264.0",
|
||||||
"@patternfly/react-icons": "4.92.10",
|
"@patternfly/react-icons": "4.92.10",
|
||||||
"@patternfly/react-table": "4.108.0",
|
"@patternfly/react-table": "4.108.0",
|
||||||
"ace-builds": "^1.10.1",
|
"ace-builds": "^1.10.1",
|
||||||
@@ -22,7 +22,7 @@
|
|||||||
"has-ansi": "5.0.1",
|
"has-ansi": "5.0.1",
|
||||||
"html-entities": "2.3.2",
|
"html-entities": "2.3.2",
|
||||||
"js-yaml": "4.1.0",
|
"js-yaml": "4.1.0",
|
||||||
"luxon": "^3.0.3",
|
"luxon": "^3.1.1",
|
||||||
"prop-types": "^15.8.1",
|
"prop-types": "^15.8.1",
|
||||||
"react": "17.0.2",
|
"react": "17.0.2",
|
||||||
"react-ace": "^10.1.0",
|
"react-ace": "^10.1.0",
|
||||||
@@ -3752,13 +3752,13 @@
|
|||||||
"integrity": "sha512-uN7JgfQsyR16YHkuGRCTIcBcnyKIqKjGkB2SGk9x1XXH3yYGenL83kpAavX9Xtozqp17KppOlybJuzcKvZMrgw=="
|
"integrity": "sha512-uN7JgfQsyR16YHkuGRCTIcBcnyKIqKjGkB2SGk9x1XXH3yYGenL83kpAavX9Xtozqp17KppOlybJuzcKvZMrgw=="
|
||||||
},
|
},
|
||||||
"node_modules/@patternfly/react-core": {
|
"node_modules/@patternfly/react-core": {
|
||||||
"version": "4.250.1",
|
"version": "4.264.0",
|
||||||
"resolved": "https://registry.npmjs.org/@patternfly/react-core/-/react-core-4.250.1.tgz",
|
"resolved": "https://registry.npmjs.org/@patternfly/react-core/-/react-core-4.264.0.tgz",
|
||||||
"integrity": "sha512-vAOZPQdZzYXl/vkHnHMIt1eC3nrPDdsuuErPatkNPwmSvilXuXmWP5wxoJ36FbSNRRURkprFwx52zMmWS3iHJA==",
|
"integrity": "sha512-tK0BMWxw8nhukev40HZ6q6d02pDnjX7oyA91vHa18aakJUKBWMaerqpG4NZVMoh0tPKX3aLNj+zyCwDALFAZZw==",
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"@patternfly/react-icons": "^4.92.6",
|
"@patternfly/react-icons": "^4.93.0",
|
||||||
"@patternfly/react-styles": "^4.91.6",
|
"@patternfly/react-styles": "^4.92.0",
|
||||||
"@patternfly/react-tokens": "^4.93.6",
|
"@patternfly/react-tokens": "^4.94.0",
|
||||||
"focus-trap": "6.9.2",
|
"focus-trap": "6.9.2",
|
||||||
"react-dropzone": "9.0.0",
|
"react-dropzone": "9.0.0",
|
||||||
"tippy.js": "5.1.2",
|
"tippy.js": "5.1.2",
|
||||||
@@ -3769,6 +3769,15 @@
|
|||||||
"react-dom": "^16.8 || ^17 || ^18"
|
"react-dom": "^16.8 || ^17 || ^18"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
|
"node_modules/@patternfly/react-core/node_modules/@patternfly/react-icons": {
|
||||||
|
"version": "4.93.0",
|
||||||
|
"resolved": "https://registry.npmjs.org/@patternfly/react-icons/-/react-icons-4.93.0.tgz",
|
||||||
|
"integrity": "sha512-OH0vORVioL+HLWMEog8/3u8jsiMCeJ0pFpvRKRhy5Uk4CdAe40k1SOBvXJP6opr+O8TLbz0q3bm8Jsh/bPaCuQ==",
|
||||||
|
"peerDependencies": {
|
||||||
|
"react": "^16.8 || ^17 || ^18",
|
||||||
|
"react-dom": "^16.8 || ^17 || ^18"
|
||||||
|
}
|
||||||
|
},
|
||||||
"node_modules/@patternfly/react-core/node_modules/tslib": {
|
"node_modules/@patternfly/react-core/node_modules/tslib": {
|
||||||
"version": "2.3.1",
|
"version": "2.3.1",
|
||||||
"resolved": "https://registry.npmjs.org/tslib/-/tslib-2.3.1.tgz",
|
"resolved": "https://registry.npmjs.org/tslib/-/tslib-2.3.1.tgz",
|
||||||
@@ -3784,9 +3793,9 @@
|
|||||||
}
|
}
|
||||||
},
|
},
|
||||||
"node_modules/@patternfly/react-styles": {
|
"node_modules/@patternfly/react-styles": {
|
||||||
"version": "4.91.10",
|
"version": "4.92.0",
|
||||||
"resolved": "https://registry.npmjs.org/@patternfly/react-styles/-/react-styles-4.91.10.tgz",
|
"resolved": "https://registry.npmjs.org/@patternfly/react-styles/-/react-styles-4.92.0.tgz",
|
||||||
"integrity": "sha512-fAG4Vjp63ohiR92F4e/Gkw5q1DSSckHKqdnEF75KUpSSBORzYP0EKMpupSd6ItpQFJw3iWs3MJi3/KIAAfU1Jw=="
|
"integrity": "sha512-B/f6iyu8UEN1+wRxdC4sLIhvJeyL8SqInDXZmwOIqK8uPJ8Lze7qrbVhkkVzbMF37/oDPVa6dZH8qZFq062LEA=="
|
||||||
},
|
},
|
||||||
"node_modules/@patternfly/react-table": {
|
"node_modules/@patternfly/react-table": {
|
||||||
"version": "4.108.0",
|
"version": "4.108.0",
|
||||||
@@ -3811,9 +3820,9 @@
|
|||||||
"integrity": "sha512-d6xOpEDfsi2CZVlPQzGeux8XMwLT9hssAsaPYExaQMuYskwb+x1x7J371tWlbBdWHroy99KnVB6qIkUbs5X3UQ=="
|
"integrity": "sha512-d6xOpEDfsi2CZVlPQzGeux8XMwLT9hssAsaPYExaQMuYskwb+x1x7J371tWlbBdWHroy99KnVB6qIkUbs5X3UQ=="
|
||||||
},
|
},
|
||||||
"node_modules/@patternfly/react-tokens": {
|
"node_modules/@patternfly/react-tokens": {
|
||||||
"version": "4.93.10",
|
"version": "4.94.0",
|
||||||
"resolved": "https://registry.npmjs.org/@patternfly/react-tokens/-/react-tokens-4.93.10.tgz",
|
"resolved": "https://registry.npmjs.org/@patternfly/react-tokens/-/react-tokens-4.94.0.tgz",
|
||||||
"integrity": "sha512-F+j1irDc9M6zvY6qNtDryhbpnHz3R8ymHRdGelNHQzPTIK88YSWEnT1c9iUI+uM/iuZol7sJmO5STtg2aPIDRQ=="
|
"integrity": "sha512-fYXxUJZnzpn89K2zzHF0cSncZZVGKrohdb5f5T1wzxwU2NZPVGpvr88xhm+V2Y/fSrrTPwXcP3IIdtNOOtJdZw=="
|
||||||
},
|
},
|
||||||
"node_modules/@pmmmwh/react-refresh-webpack-plugin": {
|
"node_modules/@pmmmwh/react-refresh-webpack-plugin": {
|
||||||
"version": "0.5.4",
|
"version": "0.5.4",
|
||||||
@@ -15468,9 +15477,9 @@
|
|||||||
}
|
}
|
||||||
},
|
},
|
||||||
"node_modules/luxon": {
|
"node_modules/luxon": {
|
||||||
"version": "3.0.3",
|
"version": "3.1.1",
|
||||||
"resolved": "https://registry.npmjs.org/luxon/-/luxon-3.0.3.tgz",
|
"resolved": "https://registry.npmjs.org/luxon/-/luxon-3.1.1.tgz",
|
||||||
"integrity": "sha512-+EfHWnF+UT7GgTnq5zXg3ldnTKL2zdv7QJgsU5bjjpbH17E3qi/puMhQyJVYuCq+FRkogvB5WB6iVvUr+E4a7w==",
|
"integrity": "sha512-Ah6DloGmvseB/pX1cAmjbFvyU/pKuwQMQqz7d0yvuDlVYLTs2WeDHQMpC8tGjm1da+BriHROW/OEIT/KfYg6xw==",
|
||||||
"engines": {
|
"engines": {
|
||||||
"node": ">=12"
|
"node": ">=12"
|
||||||
}
|
}
|
||||||
@@ -25094,19 +25103,25 @@
|
|||||||
"integrity": "sha512-uN7JgfQsyR16YHkuGRCTIcBcnyKIqKjGkB2SGk9x1XXH3yYGenL83kpAavX9Xtozqp17KppOlybJuzcKvZMrgw=="
|
"integrity": "sha512-uN7JgfQsyR16YHkuGRCTIcBcnyKIqKjGkB2SGk9x1XXH3yYGenL83kpAavX9Xtozqp17KppOlybJuzcKvZMrgw=="
|
||||||
},
|
},
|
||||||
"@patternfly/react-core": {
|
"@patternfly/react-core": {
|
||||||
"version": "4.250.1",
|
"version": "4.264.0",
|
||||||
"resolved": "https://registry.npmjs.org/@patternfly/react-core/-/react-core-4.250.1.tgz",
|
"resolved": "https://registry.npmjs.org/@patternfly/react-core/-/react-core-4.264.0.tgz",
|
||||||
"integrity": "sha512-vAOZPQdZzYXl/vkHnHMIt1eC3nrPDdsuuErPatkNPwmSvilXuXmWP5wxoJ36FbSNRRURkprFwx52zMmWS3iHJA==",
|
"integrity": "sha512-tK0BMWxw8nhukev40HZ6q6d02pDnjX7oyA91vHa18aakJUKBWMaerqpG4NZVMoh0tPKX3aLNj+zyCwDALFAZZw==",
|
||||||
"requires": {
|
"requires": {
|
||||||
"@patternfly/react-icons": "^4.92.6",
|
"@patternfly/react-icons": "^4.93.0",
|
||||||
"@patternfly/react-styles": "^4.91.6",
|
"@patternfly/react-styles": "^4.92.0",
|
||||||
"@patternfly/react-tokens": "^4.93.6",
|
"@patternfly/react-tokens": "^4.94.0",
|
||||||
"focus-trap": "6.9.2",
|
"focus-trap": "6.9.2",
|
||||||
"react-dropzone": "9.0.0",
|
"react-dropzone": "9.0.0",
|
||||||
"tippy.js": "5.1.2",
|
"tippy.js": "5.1.2",
|
||||||
"tslib": "^2.0.0"
|
"tslib": "^2.0.0"
|
||||||
},
|
},
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
|
"@patternfly/react-icons": {
|
||||||
|
"version": "4.93.0",
|
||||||
|
"resolved": "https://registry.npmjs.org/@patternfly/react-icons/-/react-icons-4.93.0.tgz",
|
||||||
|
"integrity": "sha512-OH0vORVioL+HLWMEog8/3u8jsiMCeJ0pFpvRKRhy5Uk4CdAe40k1SOBvXJP6opr+O8TLbz0q3bm8Jsh/bPaCuQ==",
|
||||||
|
"requires": {}
|
||||||
|
},
|
||||||
"tslib": {
|
"tslib": {
|
||||||
"version": "2.3.1",
|
"version": "2.3.1",
|
||||||
"resolved": "https://registry.npmjs.org/tslib/-/tslib-2.3.1.tgz",
|
"resolved": "https://registry.npmjs.org/tslib/-/tslib-2.3.1.tgz",
|
||||||
@@ -25121,9 +25136,9 @@
|
|||||||
"requires": {}
|
"requires": {}
|
||||||
},
|
},
|
||||||
"@patternfly/react-styles": {
|
"@patternfly/react-styles": {
|
||||||
"version": "4.91.10",
|
"version": "4.92.0",
|
||||||
"resolved": "https://registry.npmjs.org/@patternfly/react-styles/-/react-styles-4.91.10.tgz",
|
"resolved": "https://registry.npmjs.org/@patternfly/react-styles/-/react-styles-4.92.0.tgz",
|
||||||
"integrity": "sha512-fAG4Vjp63ohiR92F4e/Gkw5q1DSSckHKqdnEF75KUpSSBORzYP0EKMpupSd6ItpQFJw3iWs3MJi3/KIAAfU1Jw=="
|
"integrity": "sha512-B/f6iyu8UEN1+wRxdC4sLIhvJeyL8SqInDXZmwOIqK8uPJ8Lze7qrbVhkkVzbMF37/oDPVa6dZH8qZFq062LEA=="
|
||||||
},
|
},
|
||||||
"@patternfly/react-table": {
|
"@patternfly/react-table": {
|
||||||
"version": "4.108.0",
|
"version": "4.108.0",
|
||||||
@@ -25146,9 +25161,9 @@
|
|||||||
}
|
}
|
||||||
},
|
},
|
||||||
"@patternfly/react-tokens": {
|
"@patternfly/react-tokens": {
|
||||||
"version": "4.93.10",
|
"version": "4.94.0",
|
||||||
"resolved": "https://registry.npmjs.org/@patternfly/react-tokens/-/react-tokens-4.93.10.tgz",
|
"resolved": "https://registry.npmjs.org/@patternfly/react-tokens/-/react-tokens-4.94.0.tgz",
|
||||||
"integrity": "sha512-F+j1irDc9M6zvY6qNtDryhbpnHz3R8ymHRdGelNHQzPTIK88YSWEnT1c9iUI+uM/iuZol7sJmO5STtg2aPIDRQ=="
|
"integrity": "sha512-fYXxUJZnzpn89K2zzHF0cSncZZVGKrohdb5f5T1wzxwU2NZPVGpvr88xhm+V2Y/fSrrTPwXcP3IIdtNOOtJdZw=="
|
||||||
},
|
},
|
||||||
"@pmmmwh/react-refresh-webpack-plugin": {
|
"@pmmmwh/react-refresh-webpack-plugin": {
|
||||||
"version": "0.5.4",
|
"version": "0.5.4",
|
||||||
@@ -34210,9 +34225,9 @@
|
|||||||
}
|
}
|
||||||
},
|
},
|
||||||
"luxon": {
|
"luxon": {
|
||||||
"version": "3.0.3",
|
"version": "3.1.1",
|
||||||
"resolved": "https://registry.npmjs.org/luxon/-/luxon-3.0.3.tgz",
|
"resolved": "https://registry.npmjs.org/luxon/-/luxon-3.1.1.tgz",
|
||||||
"integrity": "sha512-+EfHWnF+UT7GgTnq5zXg3ldnTKL2zdv7QJgsU5bjjpbH17E3qi/puMhQyJVYuCq+FRkogvB5WB6iVvUr+E4a7w=="
|
"integrity": "sha512-Ah6DloGmvseB/pX1cAmjbFvyU/pKuwQMQqz7d0yvuDlVYLTs2WeDHQMpC8tGjm1da+BriHROW/OEIT/KfYg6xw=="
|
||||||
},
|
},
|
||||||
"lz-string": {
|
"lz-string": {
|
||||||
"version": "1.4.4",
|
"version": "1.4.4",
|
||||||
|
|||||||
@@ -8,7 +8,7 @@
|
|||||||
"dependencies": {
|
"dependencies": {
|
||||||
"@lingui/react": "3.14.0",
|
"@lingui/react": "3.14.0",
|
||||||
"@patternfly/patternfly": "4.217.1",
|
"@patternfly/patternfly": "4.217.1",
|
||||||
"@patternfly/react-core": "^4.250.1",
|
"@patternfly/react-core": "^4.264.0",
|
||||||
"@patternfly/react-icons": "4.92.10",
|
"@patternfly/react-icons": "4.92.10",
|
||||||
"@patternfly/react-table": "4.108.0",
|
"@patternfly/react-table": "4.108.0",
|
||||||
"ace-builds": "^1.10.1",
|
"ace-builds": "^1.10.1",
|
||||||
@@ -22,7 +22,7 @@
|
|||||||
"has-ansi": "5.0.1",
|
"has-ansi": "5.0.1",
|
||||||
"html-entities": "2.3.2",
|
"html-entities": "2.3.2",
|
||||||
"js-yaml": "4.1.0",
|
"js-yaml": "4.1.0",
|
||||||
"luxon": "^3.0.3",
|
"luxon": "^3.1.1",
|
||||||
"prop-types": "^15.8.1",
|
"prop-types": "^15.8.1",
|
||||||
"react": "17.0.2",
|
"react": "17.0.2",
|
||||||
"react-ace": "^10.1.0",
|
"react-ace": "^10.1.0",
|
||||||
|
|||||||
@@ -20,6 +20,10 @@ class Hosts extends Base {
|
|||||||
return this.http.get(`${this.baseUrl}${id}/all_groups/`, { params });
|
return this.http.get(`${this.baseUrl}${id}/all_groups/`, { params });
|
||||||
}
|
}
|
||||||
|
|
||||||
|
readGroups(id, params) {
|
||||||
|
return this.http.get(`${this.baseUrl}${id}/groups/`, { params });
|
||||||
|
}
|
||||||
|
|
||||||
readGroupsOptions(id) {
|
readGroupsOptions(id) {
|
||||||
return this.http.options(`${this.baseUrl}${id}/groups/`);
|
return this.http.options(`${this.baseUrl}${id}/groups/`);
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -153,6 +153,10 @@ function CredentialsStep({
|
|||||||
}))}
|
}))}
|
||||||
value={selectedType && selectedType.id}
|
value={selectedType && selectedType.id}
|
||||||
onChange={(e, id) => {
|
onChange={(e, id) => {
|
||||||
|
// Reset query params when the category of credentials is changed
|
||||||
|
history.replace({
|
||||||
|
search: '',
|
||||||
|
});
|
||||||
setSelectedType(types.find((o) => o.id === parseInt(id, 10)));
|
setSelectedType(types.find((o) => o.id === parseInt(id, 10)));
|
||||||
}}
|
}}
|
||||||
/>
|
/>
|
||||||
|
|||||||
@@ -3,6 +3,7 @@ import { act } from 'react-dom/test-utils';
|
|||||||
import { Formik } from 'formik';
|
import { Formik } from 'formik';
|
||||||
import { CredentialsAPI, CredentialTypesAPI } from 'api';
|
import { CredentialsAPI, CredentialTypesAPI } from 'api';
|
||||||
import { mountWithContexts } from '../../../../testUtils/enzymeHelpers';
|
import { mountWithContexts } from '../../../../testUtils/enzymeHelpers';
|
||||||
|
import { createMemoryHistory } from 'history';
|
||||||
import CredentialsStep from './CredentialsStep';
|
import CredentialsStep from './CredentialsStep';
|
||||||
|
|
||||||
jest.mock('../../../api/models/CredentialTypes');
|
jest.mock('../../../api/models/CredentialTypes');
|
||||||
@@ -164,6 +165,41 @@ describe('CredentialsStep', () => {
|
|||||||
});
|
});
|
||||||
});
|
});
|
||||||
|
|
||||||
|
test('should reset query params (credential.page) when selected credential type is changed', async () => {
|
||||||
|
let wrapper;
|
||||||
|
const history = createMemoryHistory({
|
||||||
|
initialEntries: ['?credential.page=2'],
|
||||||
|
});
|
||||||
|
await act(async () => {
|
||||||
|
wrapper = mountWithContexts(
|
||||||
|
<Formik>
|
||||||
|
<CredentialsStep allowCredentialsWithPasswords />
|
||||||
|
</Formik>,
|
||||||
|
{
|
||||||
|
context: { router: { history } },
|
||||||
|
}
|
||||||
|
);
|
||||||
|
});
|
||||||
|
wrapper.update();
|
||||||
|
|
||||||
|
expect(CredentialsAPI.read).toHaveBeenCalledWith({
|
||||||
|
credential_type: 1,
|
||||||
|
order_by: 'name',
|
||||||
|
page: 2,
|
||||||
|
page_size: 5,
|
||||||
|
});
|
||||||
|
|
||||||
|
await act(async () => {
|
||||||
|
wrapper.find('AnsibleSelect').invoke('onChange')({}, 3);
|
||||||
|
});
|
||||||
|
expect(CredentialsAPI.read).toHaveBeenCalledWith({
|
||||||
|
credential_type: 3,
|
||||||
|
order_by: 'name',
|
||||||
|
page: 1,
|
||||||
|
page_size: 5,
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
test("error should be shown when a credential that prompts for passwords is selected on a step that doesn't allow it", async () => {
|
test("error should be shown when a credential that prompts for passwords is selected on a step that doesn't allow it", async () => {
|
||||||
let wrapper;
|
let wrapper;
|
||||||
await act(async () => {
|
await act(async () => {
|
||||||
|
|||||||
@@ -173,6 +173,10 @@ function MultiCredentialsLookup({
|
|||||||
}))}
|
}))}
|
||||||
value={selectedType && selectedType.id}
|
value={selectedType && selectedType.id}
|
||||||
onChange={(e, id) => {
|
onChange={(e, id) => {
|
||||||
|
// Reset query params when the category of credentials is changed
|
||||||
|
history.replace({
|
||||||
|
search: '',
|
||||||
|
});
|
||||||
setSelectedType(
|
setSelectedType(
|
||||||
credentialTypes.find((o) => o.id === parseInt(id, 10))
|
credentialTypes.find((o) => o.id === parseInt(id, 10))
|
||||||
);
|
);
|
||||||
|
|||||||
@@ -6,6 +6,7 @@ import {
|
|||||||
mountWithContexts,
|
mountWithContexts,
|
||||||
waitForElement,
|
waitForElement,
|
||||||
} from '../../../testUtils/enzymeHelpers';
|
} from '../../../testUtils/enzymeHelpers';
|
||||||
|
import { createMemoryHistory } from 'history';
|
||||||
import MultiCredentialsLookup from './MultiCredentialsLookup';
|
import MultiCredentialsLookup from './MultiCredentialsLookup';
|
||||||
|
|
||||||
jest.mock('../../api');
|
jest.mock('../../api');
|
||||||
@@ -228,6 +229,53 @@ describe('<Formik><MultiCredentialsLookup /></Formik>', () => {
|
|||||||
]);
|
]);
|
||||||
});
|
});
|
||||||
|
|
||||||
|
test('should reset query params (credentials.page) when selected credential type is changed', async () => {
|
||||||
|
const history = createMemoryHistory({
|
||||||
|
initialEntries: ['?credentials.page=2'],
|
||||||
|
});
|
||||||
|
await act(async () => {
|
||||||
|
wrapper = mountWithContexts(
|
||||||
|
<Formik>
|
||||||
|
<MultiCredentialsLookup
|
||||||
|
value={credentials}
|
||||||
|
tooltip="This is credentials look up"
|
||||||
|
onChange={() => {}}
|
||||||
|
onError={() => {}}
|
||||||
|
/>
|
||||||
|
</Formik>,
|
||||||
|
{
|
||||||
|
context: { router: { history } },
|
||||||
|
}
|
||||||
|
);
|
||||||
|
});
|
||||||
|
const searchButton = await waitForElement(
|
||||||
|
wrapper,
|
||||||
|
'Button[aria-label="Search"]'
|
||||||
|
);
|
||||||
|
await act(async () => {
|
||||||
|
searchButton.invoke('onClick')();
|
||||||
|
});
|
||||||
|
expect(CredentialsAPI.read).toHaveBeenCalledWith({
|
||||||
|
credential_type: 400,
|
||||||
|
order_by: 'name',
|
||||||
|
page: 2,
|
||||||
|
page_size: 5,
|
||||||
|
});
|
||||||
|
|
||||||
|
const select = await waitForElement(wrapper, 'AnsibleSelect');
|
||||||
|
await act(async () => {
|
||||||
|
select.invoke('onChange')({}, 500);
|
||||||
|
});
|
||||||
|
wrapper.update();
|
||||||
|
|
||||||
|
expect(CredentialsAPI.read).toHaveBeenCalledWith({
|
||||||
|
credential_type: 500,
|
||||||
|
order_by: 'name',
|
||||||
|
page: 1,
|
||||||
|
page_size: 5,
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
test('should only add 1 credential per credential type except vault(see below)', async () => {
|
test('should only add 1 credential per credential type except vault(see below)', async () => {
|
||||||
const onChange = jest.fn();
|
const onChange = jest.fn();
|
||||||
await act(async () => {
|
await act(async () => {
|
||||||
|
|||||||
@@ -55,7 +55,6 @@ function DateTimePicker({ dateFieldName, timeFieldName, label }) {
|
|||||||
onChange={onDateChange}
|
onChange={onDateChange}
|
||||||
/>
|
/>
|
||||||
<TimePicker
|
<TimePicker
|
||||||
placeholder="hh:mm AM/PM"
|
|
||||||
stepMinutes={15}
|
stepMinutes={15}
|
||||||
aria-label={
|
aria-label={
|
||||||
timeFieldName.startsWith('start') ? t`Start time` : t`End time`
|
timeFieldName.startsWith('start') ? t`Start time` : t`End time`
|
||||||
|
|||||||
93
awx/ui/src/components/Schedule/shared/FrequenciesList.js
Normal file
93
awx/ui/src/components/Schedule/shared/FrequenciesList.js
Normal file
@@ -0,0 +1,93 @@
|
|||||||
|
import React, { useState } from 'react';
|
||||||
|
import { t } from '@lingui/macro';
|
||||||
|
import {
|
||||||
|
Button,
|
||||||
|
Switch,
|
||||||
|
Toolbar,
|
||||||
|
ToolbarContent,
|
||||||
|
ToolbarItem,
|
||||||
|
} from '@patternfly/react-core';
|
||||||
|
import { PencilAltIcon } from '@patternfly/react-icons';
|
||||||
|
import {
|
||||||
|
TableComposable,
|
||||||
|
Tbody,
|
||||||
|
Thead,
|
||||||
|
Th,
|
||||||
|
Tr,
|
||||||
|
Td,
|
||||||
|
} from '@patternfly/react-table';
|
||||||
|
|
||||||
|
import { useField } from 'formik';
|
||||||
|
import ContentEmpty from 'components/ContentEmpty';
|
||||||
|
|
||||||
|
function FrequenciesList({ openWizard }) {
|
||||||
|
const [isShowingRules, setIsShowingRules] = useState(true);
|
||||||
|
const [frequencies] = useField('frequencies');
|
||||||
|
const list = (freq) => (
|
||||||
|
<Tr key={freq.rrule}>
|
||||||
|
<Td>{freq.frequency}</Td>
|
||||||
|
<Td>{freq.rrule}</Td>
|
||||||
|
<Td>{t`End`}</Td>
|
||||||
|
<Td>
|
||||||
|
<Button
|
||||||
|
variant="plain"
|
||||||
|
aria-label={t`Click to toggle default value`}
|
||||||
|
ouiaId={freq ? `${freq}-button` : 'new-freq-button'}
|
||||||
|
onClick={() => {
|
||||||
|
openWizard(true);
|
||||||
|
}}
|
||||||
|
>
|
||||||
|
<PencilAltIcon />
|
||||||
|
</Button>
|
||||||
|
</Td>
|
||||||
|
</Tr>
|
||||||
|
);
|
||||||
|
return (
|
||||||
|
<>
|
||||||
|
<Toolbar>
|
||||||
|
<ToolbarContent>
|
||||||
|
<ToolbarItem>
|
||||||
|
<Button
|
||||||
|
onClick={() => {
|
||||||
|
openWizard(true);
|
||||||
|
}}
|
||||||
|
variant="secondary"
|
||||||
|
>
|
||||||
|
{isShowingRules ? t`Add RRules` : t`Add Exception`}
|
||||||
|
</Button>
|
||||||
|
</ToolbarItem>
|
||||||
|
<ToolbarItem>
|
||||||
|
<Switch
|
||||||
|
label={t`Occurances`}
|
||||||
|
labelOff={t`Exceptions`}
|
||||||
|
isChecked={isShowingRules}
|
||||||
|
onChange={(isChecked) => {
|
||||||
|
setIsShowingRules(isChecked);
|
||||||
|
}}
|
||||||
|
/>
|
||||||
|
</ToolbarItem>
|
||||||
|
</ToolbarContent>
|
||||||
|
</Toolbar>
|
||||||
|
<div css="overflow: auto">
|
||||||
|
{frequencies.value[0].frequency === '' &&
|
||||||
|
frequencies.value.length < 2 ? (
|
||||||
|
<ContentEmpty title={t`RRules`} message={t`Add RRules`} />
|
||||||
|
) : (
|
||||||
|
<TableComposable aria-label={t`RRules`} ouiaId="rrules-list">
|
||||||
|
<Thead>
|
||||||
|
<Tr>
|
||||||
|
<Th>{t`Frequency`}</Th>
|
||||||
|
<Th>{t`RRule`}</Th>
|
||||||
|
<Th>{t`Ending`}</Th>
|
||||||
|
<Th>{t`Actions`}</Th>
|
||||||
|
</Tr>
|
||||||
|
</Thead>
|
||||||
|
<Tbody>{frequencies.value.map((freq, i) => list(freq, i))}</Tbody>
|
||||||
|
</TableComposable>
|
||||||
|
)}
|
||||||
|
</div>
|
||||||
|
</>
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
export default FrequenciesList;
|
||||||
@@ -1,568 +0,0 @@
|
|||||||
import 'styled-components/macro';
|
|
||||||
import React from 'react';
|
|
||||||
import styled from 'styled-components';
|
|
||||||
import { useField } from 'formik';
|
|
||||||
|
|
||||||
import { t, Trans, Plural } from '@lingui/macro';
|
|
||||||
import { RRule } from 'rrule';
|
|
||||||
import {
|
|
||||||
Checkbox as _Checkbox,
|
|
||||||
FormGroup,
|
|
||||||
Radio,
|
|
||||||
TextInput,
|
|
||||||
} from '@patternfly/react-core';
|
|
||||||
import { required, requiredPositiveInteger } from 'util/validators';
|
|
||||||
import AnsibleSelect from '../../AnsibleSelect';
|
|
||||||
import FormField from '../../FormField';
|
|
||||||
import DateTimePicker from './DateTimePicker';
|
|
||||||
|
|
||||||
const RunOnRadio = styled(Radio)`
|
|
||||||
display: flex;
|
|
||||||
align-items: center;
|
|
||||||
|
|
||||||
label {
|
|
||||||
display: block;
|
|
||||||
width: 100%;
|
|
||||||
}
|
|
||||||
|
|
||||||
:not(:last-of-type) {
|
|
||||||
margin-bottom: 10px;
|
|
||||||
}
|
|
||||||
|
|
||||||
select:not(:first-of-type) {
|
|
||||||
margin-left: 10px;
|
|
||||||
}
|
|
||||||
`;
|
|
||||||
|
|
||||||
const RunEveryLabel = styled.p`
|
|
||||||
display: flex;
|
|
||||||
align-items: center;
|
|
||||||
`;
|
|
||||||
|
|
||||||
const Checkbox = styled(_Checkbox)`
|
|
||||||
:not(:last-of-type) {
|
|
||||||
margin-right: 10px;
|
|
||||||
}
|
|
||||||
`;
|
|
||||||
|
|
||||||
const FrequencyDetailSubform = ({ frequency, prefix, isException }) => {
|
|
||||||
const id = prefix.replace('.', '-');
|
|
||||||
const [runOnDayMonth] = useField({
|
|
||||||
name: `${prefix}.runOnDayMonth`,
|
|
||||||
});
|
|
||||||
const [runOnDayNumber] = useField({
|
|
||||||
name: `${prefix}.runOnDayNumber`,
|
|
||||||
});
|
|
||||||
const [runOnTheOccurrence] = useField({
|
|
||||||
name: `${prefix}.runOnTheOccurrence`,
|
|
||||||
});
|
|
||||||
const [runOnTheDay] = useField({
|
|
||||||
name: `${prefix}.runOnTheDay`,
|
|
||||||
});
|
|
||||||
const [runOnTheMonth] = useField({
|
|
||||||
name: `${prefix}.runOnTheMonth`,
|
|
||||||
});
|
|
||||||
const [startDate] = useField(`${prefix}.startDate`);
|
|
||||||
|
|
||||||
const [daysOfWeek, daysOfWeekMeta, daysOfWeekHelpers] = useField({
|
|
||||||
name: `${prefix}.daysOfWeek`,
|
|
||||||
validate: (val) => {
|
|
||||||
if (frequency === 'week') {
|
|
||||||
return required(t`Select a value for this field`)(val?.length > 0);
|
|
||||||
}
|
|
||||||
return undefined;
|
|
||||||
},
|
|
||||||
});
|
|
||||||
const [end, endMeta] = useField({
|
|
||||||
name: `${prefix}.end`,
|
|
||||||
validate: required(t`Select a value for this field`),
|
|
||||||
});
|
|
||||||
const [interval, intervalMeta] = useField({
|
|
||||||
name: `${prefix}.interval`,
|
|
||||||
validate: requiredPositiveInteger(),
|
|
||||||
});
|
|
||||||
const [runOn, runOnMeta] = useField({
|
|
||||||
name: `${prefix}.runOn`,
|
|
||||||
validate: (val) => {
|
|
||||||
if (frequency === 'month' || frequency === 'year') {
|
|
||||||
return required(t`Select a value for this field`)(val);
|
|
||||||
}
|
|
||||||
return undefined;
|
|
||||||
},
|
|
||||||
});
|
|
||||||
|
|
||||||
const monthOptions = [
|
|
||||||
{
|
|
||||||
key: 'january',
|
|
||||||
value: 1,
|
|
||||||
label: t`January`,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
key: 'february',
|
|
||||||
value: 2,
|
|
||||||
label: t`February`,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
key: 'march',
|
|
||||||
value: 3,
|
|
||||||
label: t`March`,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
key: 'april',
|
|
||||||
value: 4,
|
|
||||||
label: t`April`,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
key: 'may',
|
|
||||||
value: 5,
|
|
||||||
label: t`May`,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
key: 'june',
|
|
||||||
value: 6,
|
|
||||||
label: t`June`,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
key: 'july',
|
|
||||||
value: 7,
|
|
||||||
label: t`July`,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
key: 'august',
|
|
||||||
value: 8,
|
|
||||||
label: t`August`,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
key: 'september',
|
|
||||||
value: 9,
|
|
||||||
label: t`September`,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
key: 'october',
|
|
||||||
value: 10,
|
|
||||||
label: t`October`,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
key: 'november',
|
|
||||||
value: 11,
|
|
||||||
label: t`November`,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
key: 'december',
|
|
||||||
value: 12,
|
|
||||||
label: t`December`,
|
|
||||||
},
|
|
||||||
];
|
|
||||||
|
|
||||||
const updateDaysOfWeek = (day, checked) => {
|
|
||||||
const newDaysOfWeek = daysOfWeek.value ? [...daysOfWeek.value] : [];
|
|
||||||
daysOfWeekHelpers.setTouched(true);
|
|
||||||
if (checked) {
|
|
||||||
newDaysOfWeek.push(day);
|
|
||||||
daysOfWeekHelpers.setValue(newDaysOfWeek);
|
|
||||||
} else {
|
|
||||||
daysOfWeekHelpers.setValue(
|
|
||||||
newDaysOfWeek.filter((selectedDay) => selectedDay !== day)
|
|
||||||
);
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
const getPeriodLabel = () => {
|
|
||||||
switch (frequency) {
|
|
||||||
case 'minute':
|
|
||||||
return t`Minute`;
|
|
||||||
case 'hour':
|
|
||||||
return t`Hour`;
|
|
||||||
case 'day':
|
|
||||||
return t`Day`;
|
|
||||||
case 'week':
|
|
||||||
return t`Week`;
|
|
||||||
case 'month':
|
|
||||||
return t`Month`;
|
|
||||||
case 'year':
|
|
||||||
return t`Year`;
|
|
||||||
default:
|
|
||||||
throw new Error(t`Frequency did not match an expected value`);
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
const getRunEveryLabel = () => {
|
|
||||||
const intervalValue = interval.value;
|
|
||||||
|
|
||||||
switch (frequency) {
|
|
||||||
case 'minute':
|
|
||||||
return <Plural value={intervalValue} one="minute" other="minutes" />;
|
|
||||||
case 'hour':
|
|
||||||
return <Plural value={intervalValue} one="hour" other="hours" />;
|
|
||||||
case 'day':
|
|
||||||
return <Plural value={intervalValue} one="day" other="days" />;
|
|
||||||
case 'week':
|
|
||||||
return <Plural value={intervalValue} one="week" other="weeks" />;
|
|
||||||
case 'month':
|
|
||||||
return <Plural value={intervalValue} one="month" other="months" />;
|
|
||||||
case 'year':
|
|
||||||
return <Plural value={intervalValue} one="year" other="years" />;
|
|
||||||
default:
|
|
||||||
throw new Error(t`Frequency did not match an expected value`);
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
return (
|
|
||||||
<>
|
|
||||||
<p css="grid-column: 1/-1">
|
|
||||||
<b>{getPeriodLabel()}</b>
|
|
||||||
</p>
|
|
||||||
<FormGroup
|
|
||||||
name={`${prefix}.interval`}
|
|
||||||
fieldId={`schedule-run-every-${id}`}
|
|
||||||
helperTextInvalid={intervalMeta.error}
|
|
||||||
isRequired
|
|
||||||
validated={
|
|
||||||
!intervalMeta.touched || !intervalMeta.error ? 'default' : 'error'
|
|
||||||
}
|
|
||||||
label={isException ? t`Skip every` : t`Run every`}
|
|
||||||
>
|
|
||||||
<div css="display: flex">
|
|
||||||
<TextInput
|
|
||||||
css="margin-right: 10px;"
|
|
||||||
id={`schedule-run-every-${id}`}
|
|
||||||
type="number"
|
|
||||||
min="1"
|
|
||||||
step="1"
|
|
||||||
{...interval}
|
|
||||||
onChange={(value, event) => {
|
|
||||||
interval.onChange(event);
|
|
||||||
}}
|
|
||||||
/>
|
|
||||||
<RunEveryLabel>{getRunEveryLabel()}</RunEveryLabel>
|
|
||||||
</div>
|
|
||||||
</FormGroup>
|
|
||||||
{frequency === 'week' && (
|
|
||||||
<FormGroup
|
|
||||||
name={`${prefix}.daysOfWeek`}
|
|
||||||
fieldId={`schedule-days-of-week-${id}`}
|
|
||||||
helperTextInvalid={daysOfWeekMeta.error}
|
|
||||||
isRequired
|
|
||||||
validated={
|
|
||||||
!daysOfWeekMeta.touched || !daysOfWeekMeta.error
|
|
||||||
? 'default'
|
|
||||||
: 'error'
|
|
||||||
}
|
|
||||||
label={t`On days`}
|
|
||||||
>
|
|
||||||
<div css="display: flex">
|
|
||||||
<Checkbox
|
|
||||||
label={t`Sun`}
|
|
||||||
isChecked={daysOfWeek.value?.includes(RRule.SU)}
|
|
||||||
onChange={(checked) => {
|
|
||||||
updateDaysOfWeek(RRule.SU, checked);
|
|
||||||
}}
|
|
||||||
aria-label={t`Sunday`}
|
|
||||||
id={`schedule-days-of-week-sun-${id}`}
|
|
||||||
ouiaId={`schedule-days-of-week-sun-${id}`}
|
|
||||||
name={`${prefix}.daysOfWeek`}
|
|
||||||
/>
|
|
||||||
<Checkbox
|
|
||||||
label={t`Mon`}
|
|
||||||
isChecked={daysOfWeek.value?.includes(RRule.MO)}
|
|
||||||
onChange={(checked) => {
|
|
||||||
updateDaysOfWeek(RRule.MO, checked);
|
|
||||||
}}
|
|
||||||
aria-label={t`Monday`}
|
|
||||||
id={`schedule-days-of-week-mon-${id}`}
|
|
||||||
ouiaId={`schedule-days-of-week-mon-${id}`}
|
|
||||||
name={`${prefix}.daysOfWeek`}
|
|
||||||
/>
|
|
||||||
<Checkbox
|
|
||||||
label={t`Tue`}
|
|
||||||
isChecked={daysOfWeek.value?.includes(RRule.TU)}
|
|
||||||
onChange={(checked) => {
|
|
||||||
updateDaysOfWeek(RRule.TU, checked);
|
|
||||||
}}
|
|
||||||
aria-label={t`Tuesday`}
|
|
||||||
id={`schedule-days-of-week-tue-${id}`}
|
|
||||||
ouiaId={`schedule-days-of-week-tue-${id}`}
|
|
||||||
name={`${prefix}.daysOfWeek`}
|
|
||||||
/>
|
|
||||||
<Checkbox
|
|
||||||
label={t`Wed`}
|
|
||||||
isChecked={daysOfWeek.value?.includes(RRule.WE)}
|
|
||||||
onChange={(checked) => {
|
|
||||||
updateDaysOfWeek(RRule.WE, checked);
|
|
||||||
}}
|
|
||||||
aria-label={t`Wednesday`}
|
|
||||||
id={`schedule-days-of-week-wed-${id}`}
|
|
||||||
ouiaId={`schedule-days-of-week-wed-${id}`}
|
|
||||||
name={`${prefix}.daysOfWeek`}
|
|
||||||
/>
|
|
||||||
<Checkbox
|
|
||||||
label={t`Thu`}
|
|
||||||
isChecked={daysOfWeek.value?.includes(RRule.TH)}
|
|
||||||
onChange={(checked) => {
|
|
||||||
updateDaysOfWeek(RRule.TH, checked);
|
|
||||||
}}
|
|
||||||
aria-label={t`Thursday`}
|
|
||||||
id={`schedule-days-of-week-thu-${id}`}
|
|
||||||
ouiaId={`schedule-days-of-week-thu-${id}`}
|
|
||||||
name={`${prefix}.daysOfWeek`}
|
|
||||||
/>
|
|
||||||
<Checkbox
|
|
||||||
label={t`Fri`}
|
|
||||||
isChecked={daysOfWeek.value?.includes(RRule.FR)}
|
|
||||||
onChange={(checked) => {
|
|
||||||
updateDaysOfWeek(RRule.FR, checked);
|
|
||||||
}}
|
|
||||||
aria-label={t`Friday`}
|
|
||||||
id={`schedule-days-of-week-fri-${id}`}
|
|
||||||
ouiaId={`schedule-days-of-week-fri-${id}`}
|
|
||||||
name={`${prefix}.daysOfWeek`}
|
|
||||||
/>
|
|
||||||
<Checkbox
|
|
||||||
label={t`Sat`}
|
|
||||||
isChecked={daysOfWeek.value?.includes(RRule.SA)}
|
|
||||||
onChange={(checked) => {
|
|
||||||
updateDaysOfWeek(RRule.SA, checked);
|
|
||||||
}}
|
|
||||||
aria-label={t`Saturday`}
|
|
||||||
id={`schedule-days-of-week-sat-${id}`}
|
|
||||||
ouiaId={`schedule-days-of-week-sat-${id}`}
|
|
||||||
name={`${prefix}.daysOfWeek`}
|
|
||||||
/>
|
|
||||||
</div>
|
|
||||||
</FormGroup>
|
|
||||||
)}
|
|
||||||
{(frequency === 'month' || frequency === 'year') &&
|
|
||||||
!Number.isNaN(new Date(startDate.value)) && (
|
|
||||||
<FormGroup
|
|
||||||
name={`${prefix}.runOn`}
|
|
||||||
fieldId={`schedule-run-on-${id}`}
|
|
||||||
helperTextInvalid={runOnMeta.error}
|
|
||||||
isRequired
|
|
||||||
validated={
|
|
||||||
!runOnMeta.touched || !runOnMeta.error ? 'default' : 'error'
|
|
||||||
}
|
|
||||||
label={t`Run on`}
|
|
||||||
>
|
|
||||||
<RunOnRadio
|
|
||||||
id={`schedule-run-on-day-${id}`}
|
|
||||||
name={`${prefix}.runOn`}
|
|
||||||
label={
|
|
||||||
<div css="display: flex;align-items: center;">
|
|
||||||
{frequency === 'month' && (
|
|
||||||
<span
|
|
||||||
id="radio-schedule-run-on-day"
|
|
||||||
css="margin-right: 10px;"
|
|
||||||
>
|
|
||||||
<Trans>Day</Trans>
|
|
||||||
</span>
|
|
||||||
)}
|
|
||||||
{frequency === 'year' && (
|
|
||||||
<AnsibleSelect
|
|
||||||
id={`schedule-run-on-day-month-${id}`}
|
|
||||||
css="margin-right: 10px"
|
|
||||||
isDisabled={runOn.value !== 'day'}
|
|
||||||
data={monthOptions}
|
|
||||||
{...runOnDayMonth}
|
|
||||||
/>
|
|
||||||
)}
|
|
||||||
<TextInput
|
|
||||||
id={`schedule-run-on-day-number-${id}`}
|
|
||||||
type="number"
|
|
||||||
min="1"
|
|
||||||
max="31"
|
|
||||||
step="1"
|
|
||||||
isDisabled={runOn.value !== 'day'}
|
|
||||||
{...runOnDayNumber}
|
|
||||||
onChange={(value, event) => {
|
|
||||||
runOnDayNumber.onChange(event);
|
|
||||||
}}
|
|
||||||
/>
|
|
||||||
</div>
|
|
||||||
}
|
|
||||||
value="day"
|
|
||||||
isChecked={runOn.value === 'day'}
|
|
||||||
onChange={(value, event) => {
|
|
||||||
event.target.value = 'day';
|
|
||||||
runOn.onChange(event);
|
|
||||||
}}
|
|
||||||
/>
|
|
||||||
<RunOnRadio
|
|
||||||
id={`schedule-run-on-the-${id}`}
|
|
||||||
name={`${prefix}.runOn`}
|
|
||||||
label={
|
|
||||||
<div css="display: flex;align-items: center;">
|
|
||||||
<span
|
|
||||||
id={`radio-schedule-run-on-the-${id}`}
|
|
||||||
css="margin-right: 10px;"
|
|
||||||
>
|
|
||||||
<Trans>The</Trans>
|
|
||||||
</span>
|
|
||||||
<AnsibleSelect
|
|
||||||
id={`schedule-run-on-the-occurrence-${id}`}
|
|
||||||
isDisabled={runOn.value !== 'the'}
|
|
||||||
data={[
|
|
||||||
{ value: 1, key: 'first', label: t`First` },
|
|
||||||
{
|
|
||||||
value: 2,
|
|
||||||
key: 'second',
|
|
||||||
label: t`Second`,
|
|
||||||
},
|
|
||||||
{ value: 3, key: 'third', label: t`Third` },
|
|
||||||
{
|
|
||||||
value: 4,
|
|
||||||
key: 'fourth',
|
|
||||||
label: t`Fourth`,
|
|
||||||
},
|
|
||||||
{ value: 5, key: 'fifth', label: t`Fifth` },
|
|
||||||
{ value: -1, key: 'last', label: t`Last` },
|
|
||||||
]}
|
|
||||||
{...runOnTheOccurrence}
|
|
||||||
/>
|
|
||||||
<AnsibleSelect
|
|
||||||
id={`schedule-run-on-the-day-${id}`}
|
|
||||||
isDisabled={runOn.value !== 'the'}
|
|
||||||
data={[
|
|
||||||
{
|
|
||||||
value: 'sunday',
|
|
||||||
key: 'sunday',
|
|
||||||
label: t`Sunday`,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
value: 'monday',
|
|
||||||
key: 'monday',
|
|
||||||
label: t`Monday`,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
value: 'tuesday',
|
|
||||||
key: 'tuesday',
|
|
||||||
label: t`Tuesday`,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
value: 'wednesday',
|
|
||||||
key: 'wednesday',
|
|
||||||
label: t`Wednesday`,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
value: 'thursday',
|
|
||||||
key: 'thursday',
|
|
||||||
label: t`Thursday`,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
value: 'friday',
|
|
||||||
key: 'friday',
|
|
||||||
label: t`Friday`,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
value: 'saturday',
|
|
||||||
key: 'saturday',
|
|
||||||
label: t`Saturday`,
|
|
||||||
},
|
|
||||||
{ value: 'day', key: 'day', label: t`Day` },
|
|
||||||
{
|
|
||||||
value: 'weekday',
|
|
||||||
key: 'weekday',
|
|
||||||
label: t`Weekday`,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
value: 'weekendDay',
|
|
||||||
key: 'weekendDay',
|
|
||||||
label: t`Weekend day`,
|
|
||||||
},
|
|
||||||
]}
|
|
||||||
{...runOnTheDay}
|
|
||||||
/>
|
|
||||||
{frequency === 'year' && (
|
|
||||||
<>
|
|
||||||
<span
|
|
||||||
id={`of-schedule-run-on-the-month-${id}`}
|
|
||||||
css="margin-left: 10px;"
|
|
||||||
>
|
|
||||||
<Trans>of</Trans>
|
|
||||||
</span>
|
|
||||||
<AnsibleSelect
|
|
||||||
id={`schedule-run-on-the-month-${id}`}
|
|
||||||
isDisabled={runOn.value !== 'the'}
|
|
||||||
data={monthOptions}
|
|
||||||
{...runOnTheMonth}
|
|
||||||
/>
|
|
||||||
</>
|
|
||||||
)}
|
|
||||||
</div>
|
|
||||||
}
|
|
||||||
value="the"
|
|
||||||
isChecked={runOn.value === 'the'}
|
|
||||||
onChange={(value, event) => {
|
|
||||||
event.target.value = 'the';
|
|
||||||
runOn.onChange(event);
|
|
||||||
}}
|
|
||||||
/>
|
|
||||||
</FormGroup>
|
|
||||||
)}
|
|
||||||
<FormGroup
|
|
||||||
name={`${prefix}.end`}
|
|
||||||
fieldId={`schedule-end-${id}`}
|
|
||||||
helperTextInvalid={endMeta.error}
|
|
||||||
isRequired
|
|
||||||
validated={!endMeta.touched || !endMeta.error ? 'default' : 'error'}
|
|
||||||
label={t`End`}
|
|
||||||
>
|
|
||||||
<Radio
|
|
||||||
id={`end-never-${id}`}
|
|
||||||
name={`${prefix}.end`}
|
|
||||||
label={t`Never`}
|
|
||||||
value="never"
|
|
||||||
isChecked={end.value === 'never'}
|
|
||||||
onChange={(value, event) => {
|
|
||||||
event.target.value = 'never';
|
|
||||||
end.onChange(event);
|
|
||||||
}}
|
|
||||||
ouiaId={`end-never-radio-button-${id}`}
|
|
||||||
/>
|
|
||||||
<Radio
|
|
||||||
id={`end-after-${id}`}
|
|
||||||
name={`${prefix}.end`}
|
|
||||||
label={t`After number of occurrences`}
|
|
||||||
value="after"
|
|
||||||
isChecked={end.value === 'after'}
|
|
||||||
onChange={(value, event) => {
|
|
||||||
event.target.value = 'after';
|
|
||||||
end.onChange(event);
|
|
||||||
}}
|
|
||||||
ouiaId={`end-after-radio-button-${id}`}
|
|
||||||
/>
|
|
||||||
<Radio
|
|
||||||
id={`end-on-date-${id}`}
|
|
||||||
name={`${prefix}.end`}
|
|
||||||
label={t`On date`}
|
|
||||||
value="onDate"
|
|
||||||
isChecked={end.value === 'onDate'}
|
|
||||||
onChange={(value, event) => {
|
|
||||||
event.target.value = 'onDate';
|
|
||||||
end.onChange(event);
|
|
||||||
}}
|
|
||||||
ouiaId={`end-on-radio-button-${id}`}
|
|
||||||
/>
|
|
||||||
</FormGroup>
|
|
||||||
{end?.value === 'after' && (
|
|
||||||
<FormField
|
|
||||||
id={`schedule-occurrences-${id}`}
|
|
||||||
label={t`Occurrences`}
|
|
||||||
name={`${prefix}.occurrences`}
|
|
||||||
type="number"
|
|
||||||
min="1"
|
|
||||||
step="1"
|
|
||||||
isRequired
|
|
||||||
/>
|
|
||||||
)}
|
|
||||||
{end?.value === 'onDate' && (
|
|
||||||
<DateTimePicker
|
|
||||||
dateFieldName={`${prefix}.endDate`}
|
|
||||||
timeFieldName={`${prefix}.endTime`}
|
|
||||||
label={t`End date/time`}
|
|
||||||
/>
|
|
||||||
)}
|
|
||||||
</>
|
|
||||||
);
|
|
||||||
};
|
|
||||||
|
|
||||||
export default FrequencyDetailSubform;
|
|
||||||
@@ -1,30 +1,12 @@
|
|||||||
import React, { useState } from 'react';
|
import React, { useState } from 'react';
|
||||||
import { arrayOf, string } from 'prop-types';
|
import { t } from '@lingui/macro';
|
||||||
|
import { useField } from 'formik';
|
||||||
|
import { RRule } from 'rrule';
|
||||||
import { Select, SelectOption, SelectVariant } from '@patternfly/react-core';
|
import { Select, SelectOption, SelectVariant } from '@patternfly/react-core';
|
||||||
|
|
||||||
export default function FrequencySelect({
|
export default function FrequencySelect({ id, onBlur, placeholderText }) {
|
||||||
id,
|
|
||||||
value,
|
|
||||||
onChange,
|
|
||||||
onBlur,
|
|
||||||
placeholderText,
|
|
||||||
children,
|
|
||||||
}) {
|
|
||||||
const [isOpen, setIsOpen] = useState(false);
|
const [isOpen, setIsOpen] = useState(false);
|
||||||
|
const [frequency, , frequencyHelpers] = useField('freq');
|
||||||
const onSelect = (event, selectedValue) => {
|
|
||||||
if (selectedValue === 'none') {
|
|
||||||
onChange([]);
|
|
||||||
setIsOpen(false);
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
const index = value.indexOf(selectedValue);
|
|
||||||
if (index === -1) {
|
|
||||||
onChange(value.concat(selectedValue));
|
|
||||||
} else {
|
|
||||||
onChange(value.slice(0, index).concat(value.slice(index + 1)));
|
|
||||||
}
|
|
||||||
};
|
|
||||||
|
|
||||||
const onToggle = (val) => {
|
const onToggle = (val) => {
|
||||||
if (!val) {
|
if (!val) {
|
||||||
@@ -35,21 +17,26 @@ export default function FrequencySelect({
|
|||||||
|
|
||||||
return (
|
return (
|
||||||
<Select
|
<Select
|
||||||
variant={SelectVariant.checkbox}
|
onSelect={(e, v) => {
|
||||||
onSelect={onSelect}
|
frequencyHelpers.setValue(v);
|
||||||
selections={value}
|
setIsOpen(false);
|
||||||
|
}}
|
||||||
|
selections={frequency.value}
|
||||||
placeholderText={placeholderText}
|
placeholderText={placeholderText}
|
||||||
onToggle={onToggle}
|
onToggle={onToggle}
|
||||||
|
value={frequency.value}
|
||||||
isOpen={isOpen}
|
isOpen={isOpen}
|
||||||
ouiaId={`frequency-select-${id}`}
|
ouiaId={`frequency-select-${id}`}
|
||||||
|
onBlur={() => frequencyHelpers.setTouched(true)}
|
||||||
>
|
>
|
||||||
{children}
|
<SelectOption value={RRule.MINUTELY}>{t`Minute`}</SelectOption>
|
||||||
|
<SelectOption value={RRule.HOURLY}>{t`Hour`}</SelectOption>
|
||||||
|
<SelectOption value={RRule.DAILY}>{t`Day`}</SelectOption>
|
||||||
|
<SelectOption value={RRule.WEEKLY}>{t`Week`}</SelectOption>
|
||||||
|
<SelectOption value={RRule.MONTHLY}>{t`Month`}</SelectOption>
|
||||||
|
<SelectOption value={RRule.YEARLY}>{t`Year`}</SelectOption>
|
||||||
</Select>
|
</Select>
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
FrequencySelect.propTypes = {
|
|
||||||
value: arrayOf(string).isRequired,
|
|
||||||
};
|
|
||||||
|
|
||||||
export { SelectOption, SelectVariant };
|
export { SelectOption, SelectVariant };
|
||||||
|
|||||||
77
awx/ui/src/components/Schedule/shared/MonthandYearForm.js
Normal file
77
awx/ui/src/components/Schedule/shared/MonthandYearForm.js
Normal file
@@ -0,0 +1,77 @@
|
|||||||
|
import React from 'react';
|
||||||
|
import { t } from '@lingui/macro';
|
||||||
|
import AnsibleSelect from 'components/AnsibleSelect';
|
||||||
|
import styled from 'styled-components';
|
||||||
|
import {
|
||||||
|
FormGroup,
|
||||||
|
Checkbox as _Checkbox,
|
||||||
|
Grid,
|
||||||
|
GridItem,
|
||||||
|
} from '@patternfly/react-core';
|
||||||
|
import { useField } from 'formik';
|
||||||
|
import { bysetposOptions, monthOptions } from './scheduleFormHelpers';
|
||||||
|
|
||||||
|
const GroupWrapper = styled(FormGroup)`
|
||||||
|
&& .pf-c-form__group-control {
|
||||||
|
display: flex;
|
||||||
|
padding-top: 10px;
|
||||||
|
}
|
||||||
|
&& .pf-c-form__group-label {
|
||||||
|
padding-top: 20px;
|
||||||
|
}
|
||||||
|
`;
|
||||||
|
const Checkbox = styled(_Checkbox)`
|
||||||
|
:not(:last-of-type) {
|
||||||
|
margin-right: 10px;
|
||||||
|
}
|
||||||
|
`;
|
||||||
|
function MonthandYearForm({ id }) {
|
||||||
|
const [bySetPos, , bySetPosHelpers] = useField('bysetpos');
|
||||||
|
const [byMonth, , byMonthHelpers] = useField('bymonth');
|
||||||
|
|
||||||
|
return (
|
||||||
|
<>
|
||||||
|
<GroupWrapper
|
||||||
|
fieldId={`schedule-run-on-${id}`}
|
||||||
|
label={<b>{t`Run on a specific month`}</b>}
|
||||||
|
>
|
||||||
|
<Grid hasGutter>
|
||||||
|
{monthOptions.map((month) => (
|
||||||
|
<GridItem key={month.label} span={2} rowSpan={2}>
|
||||||
|
<Checkbox
|
||||||
|
label={month.label}
|
||||||
|
isChecked={byMonth.value?.includes(month.value)}
|
||||||
|
onChange={(checked) => {
|
||||||
|
if (checked) {
|
||||||
|
byMonthHelpers.setValue([...byMonth.value, month.value]);
|
||||||
|
} else {
|
||||||
|
const removed = byMonth.value.filter(
|
||||||
|
(i) => i !== month.value
|
||||||
|
);
|
||||||
|
byMonthHelpers.setValue(removed);
|
||||||
|
}
|
||||||
|
}}
|
||||||
|
id={`bymonth-${month.label}`}
|
||||||
|
ouiaId={`bymonth-${month.label}`}
|
||||||
|
name="bymonth"
|
||||||
|
/>
|
||||||
|
</GridItem>
|
||||||
|
))}
|
||||||
|
</Grid>
|
||||||
|
</GroupWrapper>
|
||||||
|
<GroupWrapper
|
||||||
|
label={<b>{t`Run on a specific week day at monthly intervals`}</b>}
|
||||||
|
>
|
||||||
|
<AnsibleSelect
|
||||||
|
id={`schedule-run-on-the-occurrence-${id}`}
|
||||||
|
data={bysetposOptions}
|
||||||
|
{...bySetPos}
|
||||||
|
onChange={(e, v) => {
|
||||||
|
bySetPosHelpers.setValue(v);
|
||||||
|
}}
|
||||||
|
/>
|
||||||
|
</GroupWrapper>
|
||||||
|
</>
|
||||||
|
);
|
||||||
|
}
|
||||||
|
export default MonthandYearForm;
|
||||||
45
awx/ui/src/components/Schedule/shared/OrdinalDayForm.js
Normal file
45
awx/ui/src/components/Schedule/shared/OrdinalDayForm.js
Normal file
@@ -0,0 +1,45 @@
|
|||||||
|
import React from 'react';
|
||||||
|
import { t } from '@lingui/macro';
|
||||||
|
import styled from 'styled-components';
|
||||||
|
import { useField } from 'formik';
|
||||||
|
import { FormGroup, TextInput } from '@patternfly/react-core';
|
||||||
|
|
||||||
|
const GroupWrapper = styled(FormGroup)`
|
||||||
|
&& .pf-c-form__group-control {
|
||||||
|
display: flex;
|
||||||
|
padding-top: 10px;
|
||||||
|
}
|
||||||
|
&& .pf-c-form__group-label {
|
||||||
|
padding-top: 20px;
|
||||||
|
}
|
||||||
|
`;
|
||||||
|
|
||||||
|
function OrdinalDayForm() {
|
||||||
|
const [byMonthDay] = useField('bymonthday');
|
||||||
|
const [byYearDay] = useField('byyearday');
|
||||||
|
return (
|
||||||
|
<GroupWrapper
|
||||||
|
label={<b>{t`On a specific number day`}</b>}
|
||||||
|
name="ordinalDay"
|
||||||
|
>
|
||||||
|
<TextInput
|
||||||
|
placeholder={t`Run on a day of month`}
|
||||||
|
aria-label={t`Type a numbered day`}
|
||||||
|
type="number"
|
||||||
|
onChange={(value, event) => {
|
||||||
|
byMonthDay.onChange(event);
|
||||||
|
}}
|
||||||
|
/>
|
||||||
|
<TextInput
|
||||||
|
placeholder={t`Run on a day of year`}
|
||||||
|
aria-label={t`Type a numbered day`}
|
||||||
|
type="number"
|
||||||
|
onChange={(value, event) => {
|
||||||
|
byYearDay.onChange(event);
|
||||||
|
}}
|
||||||
|
/>
|
||||||
|
</GroupWrapper>
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
export default OrdinalDayForm;
|
||||||
67
awx/ui/src/components/Schedule/shared/ScheduleEndForm.js
Normal file
67
awx/ui/src/components/Schedule/shared/ScheduleEndForm.js
Normal file
@@ -0,0 +1,67 @@
|
|||||||
|
import React from 'react';
|
||||||
|
import { useField } from 'formik';
|
||||||
|
import { t } from '@lingui/macro';
|
||||||
|
import { FormGroup, Radio } from '@patternfly/react-core';
|
||||||
|
import FormField from 'components/FormField';
|
||||||
|
import DateTimePicker from './DateTimePicker';
|
||||||
|
|
||||||
|
function ScheduleEndForm() {
|
||||||
|
const [endType, , { setValue }] = useField('endingType');
|
||||||
|
const [count] = useField('count');
|
||||||
|
return (
|
||||||
|
<>
|
||||||
|
<FormGroup name="end" label={t`End`}>
|
||||||
|
<Radio
|
||||||
|
id="endNever"
|
||||||
|
name={t`Never End`}
|
||||||
|
label={t`Never`}
|
||||||
|
value="never"
|
||||||
|
isChecked={endType.value === 'never'}
|
||||||
|
onChange={() => {
|
||||||
|
setValue('never');
|
||||||
|
}}
|
||||||
|
/>
|
||||||
|
<Radio
|
||||||
|
name="Count"
|
||||||
|
id="after"
|
||||||
|
label={t`After number of occurrences`}
|
||||||
|
value="after"
|
||||||
|
isChecked={endType.value === 'after'}
|
||||||
|
onChange={() => {
|
||||||
|
setValue('after');
|
||||||
|
}}
|
||||||
|
/>
|
||||||
|
<Radio
|
||||||
|
name="End Date"
|
||||||
|
label={t`On date`}
|
||||||
|
value="onDate"
|
||||||
|
id="endDate"
|
||||||
|
isChecked={endType.value === 'onDate'}
|
||||||
|
onChange={() => {
|
||||||
|
setValue('onDate');
|
||||||
|
}}
|
||||||
|
/>
|
||||||
|
</FormGroup>
|
||||||
|
{endType.value === 'after' && (
|
||||||
|
<FormField
|
||||||
|
label={t`Occurrences`}
|
||||||
|
name="count"
|
||||||
|
type="number"
|
||||||
|
min="1"
|
||||||
|
step="1"
|
||||||
|
isRequired
|
||||||
|
{...count}
|
||||||
|
/>
|
||||||
|
)}
|
||||||
|
{endType.value === 'onDate' && (
|
||||||
|
<DateTimePicker
|
||||||
|
dateFieldName="endDate"
|
||||||
|
timeFieldName="endTime"
|
||||||
|
label={t`End date/time`}
|
||||||
|
/>
|
||||||
|
)}
|
||||||
|
</>
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
export default ScheduleEndForm;
|
||||||
@@ -18,14 +18,9 @@ import SchedulePromptableFields from './SchedulePromptableFields';
|
|||||||
import ScheduleFormFields from './ScheduleFormFields';
|
import ScheduleFormFields from './ScheduleFormFields';
|
||||||
import UnsupportedScheduleForm from './UnsupportedScheduleForm';
|
import UnsupportedScheduleForm from './UnsupportedScheduleForm';
|
||||||
import parseRuleObj, { UnsupportedRRuleError } from './parseRuleObj';
|
import parseRuleObj, { UnsupportedRRuleError } from './parseRuleObj';
|
||||||
import buildRuleObj from './buildRuleObj';
|
import ScheduleFormWizard from './ScheduleFormWizard';
|
||||||
import buildRuleSet from './buildRuleSet';
|
import FrequenciesList from './FrequenciesList';
|
||||||
|
// import { validateSchedule } from './scheduleFormHelpers';
|
||||||
const NUM_DAYS_PER_FREQUENCY = {
|
|
||||||
week: 7,
|
|
||||||
month: 31,
|
|
||||||
year: 365,
|
|
||||||
};
|
|
||||||
|
|
||||||
function ScheduleForm({
|
function ScheduleForm({
|
||||||
hasDaysToKeepField,
|
hasDaysToKeepField,
|
||||||
@@ -40,15 +35,16 @@ function ScheduleForm({
|
|||||||
}) {
|
}) {
|
||||||
const [isWizardOpen, setIsWizardOpen] = useState(false);
|
const [isWizardOpen, setIsWizardOpen] = useState(false);
|
||||||
const [isSaveDisabled, setIsSaveDisabled] = useState(false);
|
const [isSaveDisabled, setIsSaveDisabled] = useState(false);
|
||||||
|
const [isScheduleWizardOpen, setIsScheduleWizardOpen] = useState(false);
|
||||||
const originalLabels = useRef([]);
|
const originalLabels = useRef([]);
|
||||||
const originalInstanceGroups = useRef([]);
|
const originalInstanceGroups = useRef([]);
|
||||||
|
|
||||||
let rruleError;
|
let rruleError;
|
||||||
const now = DateTime.now();
|
const now = DateTime.now();
|
||||||
|
|
||||||
const closestQuarterHour = DateTime.fromMillis(
|
const closestQuarterHour = DateTime.fromMillis(
|
||||||
Math.ceil(now.ts / 900000) * 900000
|
Math.ceil(now.ts / 900000) * 900000
|
||||||
);
|
);
|
||||||
const tomorrow = closestQuarterHour.plus({ days: 1 });
|
|
||||||
const isTemplate =
|
const isTemplate =
|
||||||
resource.type === 'workflow_job_template' ||
|
resource.type === 'workflow_job_template' ||
|
||||||
resource.type === 'job_template';
|
resource.type === 'job_template';
|
||||||
@@ -283,69 +279,10 @@ function ScheduleForm({
|
|||||||
}
|
}
|
||||||
const [currentDate, time] = dateToInputDateTime(closestQuarterHour.toISO());
|
const [currentDate, time] = dateToInputDateTime(closestQuarterHour.toISO());
|
||||||
|
|
||||||
const [tomorrowDate] = dateToInputDateTime(tomorrow.toISO());
|
|
||||||
const initialFrequencyOptions = {
|
|
||||||
minute: {
|
|
||||||
interval: 1,
|
|
||||||
end: 'never',
|
|
||||||
occurrences: 1,
|
|
||||||
endDate: tomorrowDate,
|
|
||||||
endTime: time,
|
|
||||||
},
|
|
||||||
hour: {
|
|
||||||
interval: 1,
|
|
||||||
end: 'never',
|
|
||||||
occurrences: 1,
|
|
||||||
endDate: tomorrowDate,
|
|
||||||
endTime: time,
|
|
||||||
},
|
|
||||||
day: {
|
|
||||||
interval: 1,
|
|
||||||
end: 'never',
|
|
||||||
occurrences: 1,
|
|
||||||
endDate: tomorrowDate,
|
|
||||||
endTime: time,
|
|
||||||
},
|
|
||||||
week: {
|
|
||||||
interval: 1,
|
|
||||||
end: 'never',
|
|
||||||
occurrences: 1,
|
|
||||||
endDate: tomorrowDate,
|
|
||||||
endTime: time,
|
|
||||||
daysOfWeek: [],
|
|
||||||
},
|
|
||||||
month: {
|
|
||||||
interval: 1,
|
|
||||||
end: 'never',
|
|
||||||
occurrences: 1,
|
|
||||||
endDate: tomorrowDate,
|
|
||||||
endTime: time,
|
|
||||||
runOn: 'day',
|
|
||||||
runOnTheOccurrence: 1,
|
|
||||||
runOnTheDay: 'sunday',
|
|
||||||
runOnDayNumber: 1,
|
|
||||||
},
|
|
||||||
year: {
|
|
||||||
interval: 1,
|
|
||||||
end: 'never',
|
|
||||||
occurrences: 1,
|
|
||||||
endDate: tomorrowDate,
|
|
||||||
endTime: time,
|
|
||||||
runOn: 'day',
|
|
||||||
runOnTheOccurrence: 1,
|
|
||||||
runOnTheDay: 'sunday',
|
|
||||||
runOnTheMonth: 1,
|
|
||||||
runOnDayMonth: 1,
|
|
||||||
runOnDayNumber: 1,
|
|
||||||
},
|
|
||||||
};
|
|
||||||
|
|
||||||
const initialValues = {
|
const initialValues = {
|
||||||
description: schedule.description || '',
|
description: schedule.description || '',
|
||||||
frequency: [],
|
frequencies: [],
|
||||||
exceptionFrequency: [],
|
exceptionFrequency: [],
|
||||||
frequencyOptions: initialFrequencyOptions,
|
|
||||||
exceptionOptions: initialFrequencyOptions,
|
|
||||||
name: schedule.name || '',
|
name: schedule.name || '',
|
||||||
startDate: currentDate,
|
startDate: currentDate,
|
||||||
startTime: time,
|
startTime: time,
|
||||||
@@ -367,11 +304,9 @@ function ScheduleForm({
|
|||||||
}
|
}
|
||||||
initialValues.daysToKeep = initialDaysToKeep;
|
initialValues.daysToKeep = initialDaysToKeep;
|
||||||
}
|
}
|
||||||
|
|
||||||
let overriddenValues = {};
|
|
||||||
if (schedule.rrule) {
|
if (schedule.rrule) {
|
||||||
try {
|
try {
|
||||||
overriddenValues = parseRuleObj(schedule);
|
parseRuleObj(schedule);
|
||||||
} catch (error) {
|
} catch (error) {
|
||||||
if (error instanceof UnsupportedRRuleError) {
|
if (error instanceof UnsupportedRRuleError) {
|
||||||
return (
|
return (
|
||||||
@@ -394,89 +329,33 @@ function ScheduleForm({
|
|||||||
if (contentLoading) {
|
if (contentLoading) {
|
||||||
return <ContentLoading />;
|
return <ContentLoading />;
|
||||||
}
|
}
|
||||||
|
const frequencies = [];
|
||||||
const validate = (values) => {
|
frequencies.push(parseRuleObj(schedule));
|
||||||
const errors = {};
|
|
||||||
|
|
||||||
values.frequency.forEach((freq) => {
|
|
||||||
const options = values.frequencyOptions[freq];
|
|
||||||
const freqErrors = {};
|
|
||||||
|
|
||||||
if (
|
|
||||||
(freq === 'month' || freq === 'year') &&
|
|
||||||
options.runOn === 'day' &&
|
|
||||||
(options.runOnDayNumber < 1 || options.runOnDayNumber > 31)
|
|
||||||
) {
|
|
||||||
freqErrors.runOn = t`Please select a day number between 1 and 31.`;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (options.end === 'after' && !options.occurrences) {
|
|
||||||
freqErrors.occurrences = t`Please enter a number of occurrences.`;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (options.end === 'onDate') {
|
|
||||||
if (
|
|
||||||
DateTime.fromFormat(
|
|
||||||
`${values.startDate} ${values.startTime}`,
|
|
||||||
'yyyy-LL-dd h:mm a'
|
|
||||||
).toMillis() >=
|
|
||||||
DateTime.fromFormat(
|
|
||||||
`${options.endDate} ${options.endTime}`,
|
|
||||||
'yyyy-LL-dd h:mm a'
|
|
||||||
).toMillis()
|
|
||||||
) {
|
|
||||||
freqErrors.endDate = t`Please select an end date/time that comes after the start date/time.`;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (
|
|
||||||
DateTime.fromISO(options.endDate)
|
|
||||||
.diff(DateTime.fromISO(values.startDate), 'days')
|
|
||||||
.toObject().days < NUM_DAYS_PER_FREQUENCY[freq]
|
|
||||||
) {
|
|
||||||
const rule = new RRule(
|
|
||||||
buildRuleObj({
|
|
||||||
startDate: values.startDate,
|
|
||||||
startTime: values.startTime,
|
|
||||||
frequency: freq,
|
|
||||||
...options,
|
|
||||||
})
|
|
||||||
);
|
|
||||||
if (rule.all().length === 0) {
|
|
||||||
errors.startDate = t`Selected date range must have at least 1 schedule occurrence.`;
|
|
||||||
freqErrors.endDate = t`Selected date range must have at least 1 schedule occurrence.`;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if (Object.keys(freqErrors).length > 0) {
|
|
||||||
if (!errors.frequencyOptions) {
|
|
||||||
errors.frequencyOptions = {};
|
|
||||||
}
|
|
||||||
errors.frequencyOptions[freq] = freqErrors;
|
|
||||||
}
|
|
||||||
});
|
|
||||||
|
|
||||||
if (values.exceptionFrequency.length > 0 && !scheduleHasInstances(values)) {
|
|
||||||
errors.exceptionFrequency = t`This schedule has no occurrences due to the selected exceptions.`;
|
|
||||||
}
|
|
||||||
|
|
||||||
return errors;
|
|
||||||
};
|
|
||||||
|
|
||||||
return (
|
return (
|
||||||
<Config>
|
<Config>
|
||||||
{() => (
|
{() => (
|
||||||
<Formik
|
<Formik
|
||||||
initialValues={{
|
initialValues={{
|
||||||
...initialValues,
|
name: schedule.name || '',
|
||||||
...overriddenValues,
|
description: schedule.description || '',
|
||||||
frequencyOptions: {
|
frequencies: frequencies || [],
|
||||||
...initialValues.frequencyOptions,
|
freq: RRule.DAILY,
|
||||||
...overriddenValues.frequencyOptions,
|
interval: 1,
|
||||||
},
|
wkst: RRule.SU,
|
||||||
exceptionOptions: {
|
byweekday: [],
|
||||||
...initialValues.exceptionOptions,
|
byweekno: [],
|
||||||
...overriddenValues.exceptionOptions,
|
bymonth: [],
|
||||||
},
|
bymonthday: '',
|
||||||
|
byyearday: '',
|
||||||
|
bysetpos: '',
|
||||||
|
until: schedule.until || null,
|
||||||
|
endDate: currentDate,
|
||||||
|
endTime: time,
|
||||||
|
count: 1,
|
||||||
|
endingType: 'never',
|
||||||
|
timezone: schedule.timezone || now.zoneName,
|
||||||
|
startDate: currentDate,
|
||||||
|
startTime: time,
|
||||||
}}
|
}}
|
||||||
onSubmit={(values) => {
|
onSubmit={(values) => {
|
||||||
submitSchedule(
|
submitSchedule(
|
||||||
@@ -488,9 +367,10 @@ function ScheduleForm({
|
|||||||
credentials
|
credentials
|
||||||
);
|
);
|
||||||
}}
|
}}
|
||||||
validate={validate}
|
validate={() => {}}
|
||||||
>
|
>
|
||||||
{(formik) => (
|
{(formik) => (
|
||||||
|
<>
|
||||||
<Form autoComplete="off" onSubmit={formik.handleSubmit}>
|
<Form autoComplete="off" onSubmit={formik.handleSubmit}>
|
||||||
<FormColumnLayout>
|
<FormColumnLayout>
|
||||||
<ScheduleFormFields
|
<ScheduleFormFields
|
||||||
@@ -517,6 +397,9 @@ function ScheduleForm({
|
|||||||
instanceGroups={originalInstanceGroups.current}
|
instanceGroups={originalInstanceGroups.current}
|
||||||
/>
|
/>
|
||||||
)}
|
)}
|
||||||
|
<FormFullWidthLayout>
|
||||||
|
<FrequenciesList openWizard={setIsScheduleWizardOpen} />
|
||||||
|
</FormFullWidthLayout>
|
||||||
<FormSubmitError error={submitError} />
|
<FormSubmitError error={submitError} />
|
||||||
<FormFullWidthLayout>
|
<FormFullWidthLayout>
|
||||||
<ActionGroup>
|
<ActionGroup>
|
||||||
@@ -531,6 +414,10 @@ function ScheduleForm({
|
|||||||
{t`Save`}
|
{t`Save`}
|
||||||
</Button>
|
</Button>
|
||||||
|
|
||||||
|
<Button
|
||||||
|
onClick={() => {}}
|
||||||
|
>{t`Preview occurances`}</Button>
|
||||||
|
|
||||||
{isTemplate && showPromptButton && (
|
{isTemplate && showPromptButton && (
|
||||||
<Button
|
<Button
|
||||||
ouiaId="schedule-form-prompt-button"
|
ouiaId="schedule-form-prompt-button"
|
||||||
@@ -555,6 +442,15 @@ function ScheduleForm({
|
|||||||
</FormFullWidthLayout>
|
</FormFullWidthLayout>
|
||||||
</FormColumnLayout>
|
</FormColumnLayout>
|
||||||
</Form>
|
</Form>
|
||||||
|
{isScheduleWizardOpen && (
|
||||||
|
<ScheduleFormWizard
|
||||||
|
staticFormFormkik={formik}
|
||||||
|
isOpen={isScheduleWizardOpen}
|
||||||
|
handleSave={() => {}}
|
||||||
|
setIsOpen={setIsScheduleWizardOpen}
|
||||||
|
/>
|
||||||
|
)}
|
||||||
|
</>
|
||||||
)}
|
)}
|
||||||
</Formik>
|
</Formik>
|
||||||
)}
|
)}
|
||||||
@@ -575,24 +471,3 @@ ScheduleForm.defaultProps = {
|
|||||||
};
|
};
|
||||||
|
|
||||||
export default ScheduleForm;
|
export default ScheduleForm;
|
||||||
|
|
||||||
function scheduleHasInstances(values) {
|
|
||||||
let rangeToCheck = 1;
|
|
||||||
values.frequency.forEach((freq) => {
|
|
||||||
if (NUM_DAYS_PER_FREQUENCY[freq] > rangeToCheck) {
|
|
||||||
rangeToCheck = NUM_DAYS_PER_FREQUENCY[freq];
|
|
||||||
}
|
|
||||||
});
|
|
||||||
|
|
||||||
const ruleSet = buildRuleSet(values, true);
|
|
||||||
const startDate = DateTime.fromISO(values.startDate);
|
|
||||||
const endDate = startDate.plus({ days: rangeToCheck });
|
|
||||||
const instances = ruleSet.between(
|
|
||||||
startDate.toJSDate(),
|
|
||||||
endDate.toJSDate(),
|
|
||||||
true,
|
|
||||||
(date, i) => i === 0
|
|
||||||
);
|
|
||||||
|
|
||||||
return instances.length > 0;
|
|
||||||
}
|
|
||||||
|
|||||||
@@ -1,41 +1,27 @@
|
|||||||
import React, { useState } from 'react';
|
import React, { useState } from 'react';
|
||||||
import { useField } from 'formik';
|
import { useField } from 'formik';
|
||||||
import { FormGroup, Title } from '@patternfly/react-core';
|
import { FormGroup } from '@patternfly/react-core';
|
||||||
import { t } from '@lingui/macro';
|
import { t } from '@lingui/macro';
|
||||||
import styled from 'styled-components';
|
|
||||||
import 'styled-components/macro';
|
|
||||||
import FormField from 'components/FormField';
|
import FormField from 'components/FormField';
|
||||||
import { required } from 'util/validators';
|
import { required } from 'util/validators';
|
||||||
import { useConfig } from 'contexts/Config';
|
import { useConfig } from 'contexts/Config';
|
||||||
import Popover from '../../Popover';
|
import Popover from '../../Popover';
|
||||||
import AnsibleSelect from '../../AnsibleSelect';
|
import AnsibleSelect from '../../AnsibleSelect';
|
||||||
import FrequencySelect, { SelectOption } from './FrequencySelect';
|
|
||||||
import getHelpText from '../../../screens/Template/shared/JobTemplate.helptext';
|
import getHelpText from '../../../screens/Template/shared/JobTemplate.helptext';
|
||||||
import { SubFormLayout, FormColumnLayout } from '../../FormLayout';
|
|
||||||
import FrequencyDetailSubform from './FrequencyDetailSubform';
|
|
||||||
import DateTimePicker from './DateTimePicker';
|
import DateTimePicker from './DateTimePicker';
|
||||||
import sortFrequencies from './sortFrequencies';
|
|
||||||
|
|
||||||
const SelectClearOption = styled(SelectOption)`
|
|
||||||
& > input[type='checkbox'] {
|
|
||||||
display: none;
|
|
||||||
}
|
|
||||||
`;
|
|
||||||
|
|
||||||
export default function ScheduleFormFields({
|
export default function ScheduleFormFields({
|
||||||
hasDaysToKeepField,
|
hasDaysToKeepField,
|
||||||
zoneOptions,
|
zoneOptions,
|
||||||
zoneLinks,
|
zoneLinks,
|
||||||
|
setTimeZone,
|
||||||
}) {
|
}) {
|
||||||
const helpText = getHelpText();
|
const helpText = getHelpText();
|
||||||
const [timezone, timezoneMeta] = useField({
|
const [timezone, timezoneMeta] = useField({
|
||||||
name: 'timezone',
|
name: 'timezone',
|
||||||
validate: required(t`Select a value for this field`),
|
validate: required(t`Select a value for this field`),
|
||||||
});
|
});
|
||||||
const [frequency, frequencyMeta, frequencyHelper] = useField({
|
|
||||||
name: 'frequency',
|
|
||||||
validate: required(t`Select a value for this field`),
|
|
||||||
});
|
|
||||||
const [timezoneMessage, setTimezoneMessage] = useState('');
|
const [timezoneMessage, setTimezoneMessage] = useState('');
|
||||||
const warnLinkedTZ = (event, selectedValue) => {
|
const warnLinkedTZ = (event, selectedValue) => {
|
||||||
if (zoneLinks[selectedValue]) {
|
if (zoneLinks[selectedValue]) {
|
||||||
@@ -46,6 +32,7 @@ export default function ScheduleFormFields({
|
|||||||
setTimezoneMessage('');
|
setTimezoneMessage('');
|
||||||
}
|
}
|
||||||
timezone.onChange(event, selectedValue);
|
timezone.onChange(event, selectedValue);
|
||||||
|
setTimeZone(zoneLinks(selectedValue));
|
||||||
};
|
};
|
||||||
let timezoneValidatedStatus = 'default';
|
let timezoneValidatedStatus = 'default';
|
||||||
if (timezoneMeta.touched && timezoneMeta.error) {
|
if (timezoneMeta.touched && timezoneMeta.error) {
|
||||||
@@ -55,16 +42,6 @@ export default function ScheduleFormFields({
|
|||||||
}
|
}
|
||||||
const config = useConfig();
|
const config = useConfig();
|
||||||
|
|
||||||
const [exceptionFrequency, exceptionFrequencyMeta, exceptionFrequencyHelper] =
|
|
||||||
useField({
|
|
||||||
name: 'exceptionFrequency',
|
|
||||||
validate: required(t`Select a value for this field`),
|
|
||||||
});
|
|
||||||
|
|
||||||
const updateFrequency = (setFrequency) => (values) => {
|
|
||||||
setFrequency(values.sort(sortFrequencies));
|
|
||||||
};
|
|
||||||
|
|
||||||
return (
|
return (
|
||||||
<>
|
<>
|
||||||
<FormField
|
<FormField
|
||||||
@@ -103,33 +80,7 @@ export default function ScheduleFormFields({
|
|||||||
onChange={warnLinkedTZ}
|
onChange={warnLinkedTZ}
|
||||||
/>
|
/>
|
||||||
</FormGroup>
|
</FormGroup>
|
||||||
<FormGroup
|
|
||||||
name="frequency"
|
|
||||||
fieldId="schedule-frequency"
|
|
||||||
helperTextInvalid={frequencyMeta.error}
|
|
||||||
validated={
|
|
||||||
!frequencyMeta.touched || !frequencyMeta.error ? 'default' : 'error'
|
|
||||||
}
|
|
||||||
label={t`Repeat frequency`}
|
|
||||||
>
|
|
||||||
<FrequencySelect
|
|
||||||
id="schedule-frequency"
|
|
||||||
onChange={updateFrequency(frequencyHelper.setValue)}
|
|
||||||
value={frequency.value}
|
|
||||||
placeholderText={
|
|
||||||
frequency.value.length ? t`Select frequency` : t`None (run once)`
|
|
||||||
}
|
|
||||||
onBlur={frequencyHelper.setTouched}
|
|
||||||
>
|
|
||||||
<SelectClearOption value="none">{t`None (run once)`}</SelectClearOption>
|
|
||||||
<SelectOption value="minute">{t`Minute`}</SelectOption>
|
|
||||||
<SelectOption value="hour">{t`Hour`}</SelectOption>
|
|
||||||
<SelectOption value="day">{t`Day`}</SelectOption>
|
|
||||||
<SelectOption value="week">{t`Week`}</SelectOption>
|
|
||||||
<SelectOption value="month">{t`Month`}</SelectOption>
|
|
||||||
<SelectOption value="year">{t`Year`}</SelectOption>
|
|
||||||
</FrequencySelect>
|
|
||||||
</FormGroup>
|
|
||||||
{hasDaysToKeepField ? (
|
{hasDaysToKeepField ? (
|
||||||
<FormField
|
<FormField
|
||||||
id="schedule-days-to-keep"
|
id="schedule-days-to-keep"
|
||||||
@@ -140,68 +91,6 @@ export default function ScheduleFormFields({
|
|||||||
isRequired
|
isRequired
|
||||||
/>
|
/>
|
||||||
) : null}
|
) : null}
|
||||||
{frequency.value.length ? (
|
|
||||||
<SubFormLayout>
|
|
||||||
<Title size="md" headingLevel="h4">
|
|
||||||
{t`Frequency Details`}
|
|
||||||
</Title>
|
|
||||||
{frequency.value.map((val) => (
|
|
||||||
<FormColumnLayout key={val} stacked>
|
|
||||||
<FrequencyDetailSubform
|
|
||||||
frequency={val}
|
|
||||||
prefix={`frequencyOptions.${val}`}
|
|
||||||
/>
|
|
||||||
</FormColumnLayout>
|
|
||||||
))}
|
|
||||||
<Title
|
|
||||||
size="md"
|
|
||||||
headingLevel="h4"
|
|
||||||
css="margin-top: var(--pf-c-card--child--PaddingRight)"
|
|
||||||
>{t`Exceptions`}</Title>
|
|
||||||
<FormColumnLayout stacked>
|
|
||||||
<FormGroup
|
|
||||||
name="exceptions"
|
|
||||||
fieldId="exception-frequency"
|
|
||||||
helperTextInvalid={exceptionFrequencyMeta.error}
|
|
||||||
validated={
|
|
||||||
!exceptionFrequencyMeta.touched || !exceptionFrequencyMeta.error
|
|
||||||
? 'default'
|
|
||||||
: 'error'
|
|
||||||
}
|
|
||||||
label={t`Add exceptions`}
|
|
||||||
>
|
|
||||||
<FrequencySelect
|
|
||||||
id="exception-frequency"
|
|
||||||
onChange={updateFrequency(exceptionFrequencyHelper.setValue)}
|
|
||||||
value={exceptionFrequency.value}
|
|
||||||
placeholderText={
|
|
||||||
exceptionFrequency.value.length
|
|
||||||
? t`Select frequency`
|
|
||||||
: t`None`
|
|
||||||
}
|
|
||||||
onBlur={exceptionFrequencyHelper.setTouched}
|
|
||||||
>
|
|
||||||
<SelectClearOption value="none">{t`None`}</SelectClearOption>
|
|
||||||
<SelectOption value="minute">{t`Minute`}</SelectOption>
|
|
||||||
<SelectOption value="hour">{t`Hour`}</SelectOption>
|
|
||||||
<SelectOption value="day">{t`Day`}</SelectOption>
|
|
||||||
<SelectOption value="week">{t`Week`}</SelectOption>
|
|
||||||
<SelectOption value="month">{t`Month`}</SelectOption>
|
|
||||||
<SelectOption value="year">{t`Year`}</SelectOption>
|
|
||||||
</FrequencySelect>
|
|
||||||
</FormGroup>
|
|
||||||
</FormColumnLayout>
|
|
||||||
{exceptionFrequency.value.map((val) => (
|
|
||||||
<FormColumnLayout key={val} stacked>
|
|
||||||
<FrequencyDetailSubform
|
|
||||||
frequency={val}
|
|
||||||
prefix={`exceptionOptions.${val}`}
|
|
||||||
isException
|
|
||||||
/>
|
|
||||||
</FormColumnLayout>
|
|
||||||
))}
|
|
||||||
</SubFormLayout>
|
|
||||||
) : null}
|
|
||||||
</>
|
</>
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|||||||
199
awx/ui/src/components/Schedule/shared/ScheduleFormWizard.js
Normal file
199
awx/ui/src/components/Schedule/shared/ScheduleFormWizard.js
Normal file
@@ -0,0 +1,199 @@
|
|||||||
|
import React from 'react';
|
||||||
|
import {
|
||||||
|
Button,
|
||||||
|
FormGroup,
|
||||||
|
TextInput,
|
||||||
|
Title,
|
||||||
|
Wizard,
|
||||||
|
WizardContextConsumer,
|
||||||
|
WizardFooter,
|
||||||
|
} from '@patternfly/react-core';
|
||||||
|
import { t } from '@lingui/macro';
|
||||||
|
import styled from 'styled-components';
|
||||||
|
import { RRule } from 'rrule';
|
||||||
|
import { useField, useFormikContext } from 'formik';
|
||||||
|
import { DateTime } from 'luxon';
|
||||||
|
import { formatDateString } from 'util/dates';
|
||||||
|
import FrequencySelect from './FrequencySelect';
|
||||||
|
import MonthandYearForm from './MonthandYearForm';
|
||||||
|
import OrdinalDayForm from './OrdinalDayForm';
|
||||||
|
import WeekdayForm from './WeekdayForm';
|
||||||
|
import ScheduleEndForm from './ScheduleEndForm';
|
||||||
|
import parseRuleObj from './parseRuleObj';
|
||||||
|
import { buildDtStartObj } from './buildRuleObj';
|
||||||
|
|
||||||
|
const GroupWrapper = styled(FormGroup)`
|
||||||
|
&& .pf-c-form__group-control {
|
||||||
|
display: flex;
|
||||||
|
padding-top: 10px;
|
||||||
|
}
|
||||||
|
&& .pf-c-form__group-label {
|
||||||
|
padding-top: 20px;
|
||||||
|
}
|
||||||
|
`;
|
||||||
|
|
||||||
|
function ScheduleFormWizard({ isOpen, setIsOpen }) {
|
||||||
|
const { values, resetForm, initialValues } = useFormikContext();
|
||||||
|
const [freq, freqMeta] = useField('freq');
|
||||||
|
const [{ value: frequenciesValue }] = useField('frequencies');
|
||||||
|
const [interval, , intervalHelpers] = useField('interval');
|
||||||
|
|
||||||
|
const handleSubmit = (goToStepById) => {
|
||||||
|
const {
|
||||||
|
name,
|
||||||
|
description,
|
||||||
|
endingType,
|
||||||
|
endTime,
|
||||||
|
endDate,
|
||||||
|
timezone,
|
||||||
|
startDate,
|
||||||
|
startTime,
|
||||||
|
frequencies,
|
||||||
|
...rest
|
||||||
|
} = values;
|
||||||
|
if (endingType === 'onDate') {
|
||||||
|
const dt = DateTime.fromFormat(
|
||||||
|
`${endDate} ${endTime}`,
|
||||||
|
'yyyy-MM-dd h:mm a',
|
||||||
|
{
|
||||||
|
zone: timezone,
|
||||||
|
}
|
||||||
|
);
|
||||||
|
rest.until = formatDateString(dt, timezone);
|
||||||
|
|
||||||
|
delete rest.count;
|
||||||
|
}
|
||||||
|
if (endingType === 'never') delete rest.count;
|
||||||
|
|
||||||
|
const rule = new RRule(rest);
|
||||||
|
|
||||||
|
const start = buildDtStartObj({
|
||||||
|
startDate: values.startDate,
|
||||||
|
startTime: values.startTime,
|
||||||
|
timezone: values.timezone,
|
||||||
|
frequency: values.freq,
|
||||||
|
});
|
||||||
|
const newFrequency = parseRuleObj({
|
||||||
|
timezone,
|
||||||
|
frequency: freq.value,
|
||||||
|
rrule: rule.toString(),
|
||||||
|
dtstart: start,
|
||||||
|
});
|
||||||
|
if (goToStepById) {
|
||||||
|
goToStepById(1);
|
||||||
|
}
|
||||||
|
|
||||||
|
resetForm({
|
||||||
|
values: {
|
||||||
|
...initialValues,
|
||||||
|
description: values.description,
|
||||||
|
name: values.name,
|
||||||
|
startDate: values.startDate,
|
||||||
|
startTime: values.startTime,
|
||||||
|
timezone: values.timezone,
|
||||||
|
frequencies: frequenciesValue[0].frequency.length
|
||||||
|
? [...frequenciesValue, newFrequency]
|
||||||
|
: [newFrequency],
|
||||||
|
},
|
||||||
|
});
|
||||||
|
};
|
||||||
|
const CustomFooter = (
|
||||||
|
<WizardFooter>
|
||||||
|
<WizardContextConsumer>
|
||||||
|
{({ activeStep, onNext, onBack, goToStepById }) => (
|
||||||
|
<>
|
||||||
|
{activeStep.id === 2 ? (
|
||||||
|
<>
|
||||||
|
<Button
|
||||||
|
variant="primary"
|
||||||
|
onClick={() => {
|
||||||
|
handleSubmit(true, goToStepById);
|
||||||
|
}}
|
||||||
|
>{t`Finish and create new`}</Button>
|
||||||
|
<Button
|
||||||
|
variant="secondary"
|
||||||
|
onClick={() => {
|
||||||
|
handleSubmit(false);
|
||||||
|
setIsOpen(false);
|
||||||
|
}}
|
||||||
|
>{t`Finish and close`}</Button>
|
||||||
|
</>
|
||||||
|
) : (
|
||||||
|
<Button variant="primary" onClick={onNext}>{t`Next`}</Button>
|
||||||
|
)}
|
||||||
|
|
||||||
|
<Button variant="secondary" onClick={onBack}>{t`Back`}</Button>
|
||||||
|
<Button
|
||||||
|
variant="plain"
|
||||||
|
onClick={() => {
|
||||||
|
setIsOpen(false);
|
||||||
|
resetForm({
|
||||||
|
values: {
|
||||||
|
...initialValues,
|
||||||
|
description: values.description,
|
||||||
|
name: values.name,
|
||||||
|
startDate: values.startDate,
|
||||||
|
startTime: values.startTime,
|
||||||
|
timezone: values.timezone,
|
||||||
|
frequencies: values.frequencies,
|
||||||
|
},
|
||||||
|
});
|
||||||
|
}}
|
||||||
|
>{t`Cancel`}</Button>
|
||||||
|
</>
|
||||||
|
)}
|
||||||
|
</WizardContextConsumer>
|
||||||
|
</WizardFooter>
|
||||||
|
);
|
||||||
|
|
||||||
|
return (
|
||||||
|
<Wizard
|
||||||
|
onClose={() => setIsOpen(false)}
|
||||||
|
isOpen={isOpen}
|
||||||
|
footer={CustomFooter}
|
||||||
|
steps={[
|
||||||
|
{
|
||||||
|
key: 'frequency',
|
||||||
|
name: 'Frequency',
|
||||||
|
id: 1,
|
||||||
|
component: (
|
||||||
|
<>
|
||||||
|
<Title size="md" headingLevel="h4">{t`Repeat frequency`}</Title>
|
||||||
|
<GroupWrapper
|
||||||
|
name="freq"
|
||||||
|
fieldId="schedule-frequency"
|
||||||
|
isRequired
|
||||||
|
helperTextInvalid={freqMeta.error}
|
||||||
|
validated={
|
||||||
|
!freqMeta.touched || !freqMeta.error ? 'default' : 'error'
|
||||||
|
}
|
||||||
|
label={<b>{t`Frequency`}</b>}
|
||||||
|
>
|
||||||
|
<FrequencySelect />
|
||||||
|
</GroupWrapper>
|
||||||
|
<GroupWrapper isRequired label={<b>{t`Interval`}</b>}>
|
||||||
|
<TextInput
|
||||||
|
type="number"
|
||||||
|
value={interval.value}
|
||||||
|
placeholder={t`Choose an interval for the schedule`}
|
||||||
|
aria-label={t`Choose an interval for the schedule`}
|
||||||
|
onChange={(v) => intervalHelpers.setValue(v)}
|
||||||
|
/>
|
||||||
|
</GroupWrapper>
|
||||||
|
<WeekdayForm />
|
||||||
|
<MonthandYearForm />
|
||||||
|
<OrdinalDayForm />
|
||||||
|
</>
|
||||||
|
),
|
||||||
|
},
|
||||||
|
{
|
||||||
|
name: 'End',
|
||||||
|
key: 'end',
|
||||||
|
id: 2,
|
||||||
|
component: <ScheduleEndForm />,
|
||||||
|
},
|
||||||
|
]}
|
||||||
|
/>
|
||||||
|
);
|
||||||
|
}
|
||||||
|
export default ScheduleFormWizard;
|
||||||
164
awx/ui/src/components/Schedule/shared/WeekdayForm.js
Normal file
164
awx/ui/src/components/Schedule/shared/WeekdayForm.js
Normal file
@@ -0,0 +1,164 @@
|
|||||||
|
import React, { useState } from 'react';
|
||||||
|
import { t } from '@lingui/macro';
|
||||||
|
import {
|
||||||
|
Checkbox as _Checkbox,
|
||||||
|
FormGroup,
|
||||||
|
Select,
|
||||||
|
SelectOption,
|
||||||
|
} from '@patternfly/react-core';
|
||||||
|
import { useField } from 'formik';
|
||||||
|
import { RRule } from 'rrule';
|
||||||
|
import styled from 'styled-components';
|
||||||
|
import { weekdayOptions } from './scheduleFormHelpers';
|
||||||
|
|
||||||
|
const Checkbox = styled(_Checkbox)`
|
||||||
|
:not(:last-of-type) {
|
||||||
|
margin-right: 10px;
|
||||||
|
}
|
||||||
|
`;
|
||||||
|
const GroupWrapper = styled(FormGroup)`
|
||||||
|
&& .pf-c-form__group-control {
|
||||||
|
display: flex;
|
||||||
|
padding-top: 10px;
|
||||||
|
}
|
||||||
|
&& .pf-c-form__group-label {
|
||||||
|
padding-top: 20px;
|
||||||
|
}
|
||||||
|
`;
|
||||||
|
|
||||||
|
function WeekdayForm({ id }) {
|
||||||
|
const [isOpen, setIsOpen] = useState(false);
|
||||||
|
const [daysOfWeek, daysOfWeekMeta, daysOfWeekHelpers] = useField('byweekday');
|
||||||
|
const [weekStartDay, , weekStartDayHelpers] = useField('wkst');
|
||||||
|
const updateDaysOfWeek = (day, checked) => {
|
||||||
|
const newDaysOfWeek = daysOfWeek.value ? [...daysOfWeek.value] : [];
|
||||||
|
daysOfWeekHelpers.setTouched(true);
|
||||||
|
|
||||||
|
if (checked) {
|
||||||
|
newDaysOfWeek.push(day);
|
||||||
|
daysOfWeekHelpers.setValue(newDaysOfWeek);
|
||||||
|
} else {
|
||||||
|
daysOfWeekHelpers.setValue(
|
||||||
|
newDaysOfWeek.filter((selectedDay) => selectedDay !== day)
|
||||||
|
);
|
||||||
|
}
|
||||||
|
};
|
||||||
|
return (
|
||||||
|
<>
|
||||||
|
<GroupWrapper
|
||||||
|
name="wkst"
|
||||||
|
label={<b>{t`Select the first day of the week`}</b>}
|
||||||
|
>
|
||||||
|
<Select
|
||||||
|
onSelect={(e, value) => {
|
||||||
|
weekStartDayHelpers.setValue(value);
|
||||||
|
setIsOpen(false);
|
||||||
|
}}
|
||||||
|
onBlur={() => setIsOpen(false)}
|
||||||
|
selections={weekStartDay.value}
|
||||||
|
onToggle={(isopen) => setIsOpen(isopen)}
|
||||||
|
isOpen={isOpen}
|
||||||
|
id={`schedule-run-on-the-day-${id}`}
|
||||||
|
onChange={(e, v) => {
|
||||||
|
weekStartDayHelpers.setValue(v);
|
||||||
|
}}
|
||||||
|
{...weekStartDay}
|
||||||
|
>
|
||||||
|
{weekdayOptions.map(({ key, value, label }) => (
|
||||||
|
<SelectOption key={key} value={value}>
|
||||||
|
{label}
|
||||||
|
</SelectOption>
|
||||||
|
))}
|
||||||
|
</Select>
|
||||||
|
</GroupWrapper>
|
||||||
|
<GroupWrapper
|
||||||
|
name="byweekday"
|
||||||
|
fieldId={`schedule-days-of-week-${id}`}
|
||||||
|
helperTextInvalid={daysOfWeekMeta.error}
|
||||||
|
validated={
|
||||||
|
!daysOfWeekMeta.touched || !daysOfWeekMeta.error ? 'default' : 'error'
|
||||||
|
}
|
||||||
|
label={<b>{t`On selected day(s) of the week`}</b>}
|
||||||
|
>
|
||||||
|
<Checkbox
|
||||||
|
label={t`Sun`}
|
||||||
|
isChecked={daysOfWeek.value?.includes(RRule.SU)}
|
||||||
|
onChange={(checked) => {
|
||||||
|
updateDaysOfWeek(RRule.SU, checked);
|
||||||
|
}}
|
||||||
|
aria-label={t`Sunday`}
|
||||||
|
id={`schedule-days-of-week-sun-${id}`}
|
||||||
|
ouiaId={`schedule-days-of-week-sun-${id}`}
|
||||||
|
name="daysOfWeek"
|
||||||
|
/>
|
||||||
|
<Checkbox
|
||||||
|
label={t`Mon`}
|
||||||
|
isChecked={daysOfWeek.value?.includes(RRule.MO)}
|
||||||
|
onChange={(checked) => {
|
||||||
|
updateDaysOfWeek(RRule.MO, checked);
|
||||||
|
}}
|
||||||
|
aria-label={t`Monday`}
|
||||||
|
id={`schedule-days-of-week-mon-${id}`}
|
||||||
|
ouiaId={`schedule-days-of-week-mon-${id}`}
|
||||||
|
name="daysOfWeek"
|
||||||
|
/>
|
||||||
|
<Checkbox
|
||||||
|
label={t`Tue`}
|
||||||
|
isChecked={daysOfWeek.value?.includes(RRule.TU)}
|
||||||
|
onChange={(checked) => {
|
||||||
|
updateDaysOfWeek(RRule.TU, checked);
|
||||||
|
}}
|
||||||
|
aria-label={t`Tuesday`}
|
||||||
|
id={`schedule-days-of-week-tue-${id}`}
|
||||||
|
ouiaId={`schedule-days-of-week-tue-${id}`}
|
||||||
|
name="daysOfWeek"
|
||||||
|
/>
|
||||||
|
<Checkbox
|
||||||
|
label={t`Wed`}
|
||||||
|
isChecked={daysOfWeek.value?.includes(RRule.WE)}
|
||||||
|
onChange={(checked) => {
|
||||||
|
updateDaysOfWeek(RRule.WE, checked);
|
||||||
|
}}
|
||||||
|
aria-label={t`Wednesday`}
|
||||||
|
id={`schedule-days-of-week-wed-${id}`}
|
||||||
|
ouiaId={`schedule-days-of-week-wed-${id}`}
|
||||||
|
name="daysOfWeek"
|
||||||
|
/>
|
||||||
|
<Checkbox
|
||||||
|
label={t`Thu`}
|
||||||
|
isChecked={daysOfWeek.value?.includes(RRule.TH)}
|
||||||
|
onChange={(checked) => {
|
||||||
|
updateDaysOfWeek(RRule.TH, checked);
|
||||||
|
}}
|
||||||
|
aria-label={t`Thursday`}
|
||||||
|
id={`schedule-days-of-week-thu-${id}`}
|
||||||
|
ouiaId={`schedule-days-of-week-thu-${id}`}
|
||||||
|
name="daysOfWeek"
|
||||||
|
/>
|
||||||
|
<Checkbox
|
||||||
|
label={t`Fri`}
|
||||||
|
isChecked={daysOfWeek.value?.includes(RRule.FR)}
|
||||||
|
onChange={(checked) => {
|
||||||
|
updateDaysOfWeek(RRule.FR, checked);
|
||||||
|
}}
|
||||||
|
aria-label={t`Friday`}
|
||||||
|
id={`schedule-days-of-week-fri-${id}`}
|
||||||
|
ouiaId={`schedule-days-of-week-fri-${id}`}
|
||||||
|
name="daysOfWeek"
|
||||||
|
/>
|
||||||
|
<Checkbox
|
||||||
|
label={t`Sat`}
|
||||||
|
isChecked={daysOfWeek.value?.includes(RRule.SA)}
|
||||||
|
onChange={(checked) => {
|
||||||
|
updateDaysOfWeek(RRule.SA, checked);
|
||||||
|
}}
|
||||||
|
aria-label={t`Saturday`}
|
||||||
|
id={`schedule-days-of-week-sat-${id}`}
|
||||||
|
ouiaId={`schedule-days-of-week-sat-${id}`}
|
||||||
|
name="daysOfWeek"
|
||||||
|
/>
|
||||||
|
</GroupWrapper>
|
||||||
|
</>
|
||||||
|
);
|
||||||
|
}
|
||||||
|
export default WeekdayForm;
|
||||||
@@ -1,7 +1,5 @@
|
|||||||
import { t } from '@lingui/macro';
|
|
||||||
import { RRule } from 'rrule';
|
import { RRule } from 'rrule';
|
||||||
import { DateTime } from 'luxon';
|
import { DateTime } from 'luxon';
|
||||||
import { getRRuleDayConstants } from 'util/dates';
|
|
||||||
|
|
||||||
window.RRule = RRule;
|
window.RRule = RRule;
|
||||||
window.DateTime = DateTime;
|
window.DateTime = DateTime;
|
||||||
@@ -22,7 +20,7 @@ export function buildDtStartObj(values) {
|
|||||||
startHour
|
startHour
|
||||||
)}${pad(startMinute)}00`;
|
)}${pad(startMinute)}00`;
|
||||||
const rruleString = values.timezone
|
const rruleString = values.timezone
|
||||||
? `DTSTART;TZID=${values.timezone}:${dateString}`
|
? `DTSTART;TZID=${values.timezone}${dateString}`
|
||||||
: `DTSTART:${dateString}Z`;
|
: `DTSTART:${dateString}Z`;
|
||||||
const rule = RRule.fromString(rruleString);
|
const rule = RRule.fromString(rruleString);
|
||||||
|
|
||||||
@@ -38,7 +36,8 @@ function pad(num) {
|
|||||||
|
|
||||||
export default function buildRuleObj(values, includeStart) {
|
export default function buildRuleObj(values, includeStart) {
|
||||||
const ruleObj = {
|
const ruleObj = {
|
||||||
interval: values.interval,
|
interval: values.interval || 1,
|
||||||
|
freq: values.freq,
|
||||||
};
|
};
|
||||||
|
|
||||||
if (includeStart) {
|
if (includeStart) {
|
||||||
@@ -49,68 +48,6 @@ export default function buildRuleObj(values, includeStart) {
|
|||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
switch (values.frequency) {
|
|
||||||
case 'none':
|
|
||||||
ruleObj.count = 1;
|
|
||||||
ruleObj.freq = RRule.MINUTELY;
|
|
||||||
break;
|
|
||||||
case 'minute':
|
|
||||||
ruleObj.freq = RRule.MINUTELY;
|
|
||||||
break;
|
|
||||||
case 'hour':
|
|
||||||
ruleObj.freq = RRule.HOURLY;
|
|
||||||
break;
|
|
||||||
case 'day':
|
|
||||||
ruleObj.freq = RRule.DAILY;
|
|
||||||
break;
|
|
||||||
case 'week':
|
|
||||||
ruleObj.freq = RRule.WEEKLY;
|
|
||||||
ruleObj.byweekday = values.daysOfWeek;
|
|
||||||
break;
|
|
||||||
case 'month':
|
|
||||||
ruleObj.freq = RRule.MONTHLY;
|
|
||||||
if (values.runOn === 'day') {
|
|
||||||
ruleObj.bymonthday = values.runOnDayNumber;
|
|
||||||
} else if (values.runOn === 'the') {
|
|
||||||
ruleObj.bysetpos = parseInt(values.runOnTheOccurrence, 10);
|
|
||||||
ruleObj.byweekday = getRRuleDayConstants(values.runOnTheDay);
|
|
||||||
}
|
|
||||||
break;
|
|
||||||
case 'year':
|
|
||||||
ruleObj.freq = RRule.YEARLY;
|
|
||||||
if (values.runOn === 'day') {
|
|
||||||
ruleObj.bymonth = parseInt(values.runOnDayMonth, 10);
|
|
||||||
ruleObj.bymonthday = values.runOnDayNumber;
|
|
||||||
} else if (values.runOn === 'the') {
|
|
||||||
ruleObj.bysetpos = parseInt(values.runOnTheOccurrence, 10);
|
|
||||||
ruleObj.byweekday = getRRuleDayConstants(values.runOnTheDay);
|
|
||||||
ruleObj.bymonth = parseInt(values.runOnTheMonth, 10);
|
|
||||||
}
|
|
||||||
break;
|
|
||||||
default:
|
|
||||||
throw new Error(t`Frequency did not match an expected value`);
|
|
||||||
}
|
|
||||||
|
|
||||||
if (values.frequency !== 'none') {
|
|
||||||
switch (values.end) {
|
|
||||||
case 'never':
|
|
||||||
break;
|
|
||||||
case 'after':
|
|
||||||
ruleObj.count = values.occurrences;
|
|
||||||
break;
|
|
||||||
case 'onDate': {
|
|
||||||
ruleObj.until = buildDateTime(
|
|
||||||
values.endDate,
|
|
||||||
values.endTime,
|
|
||||||
values.timezone
|
|
||||||
);
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
default:
|
|
||||||
throw new Error(t`End did not match an expected value (${values.end})`);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return ruleObj;
|
return ruleObj;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -1,5 +1,6 @@
|
|||||||
import { RRule, RRuleSet } from 'rrule';
|
import { RRule, RRuleSet } from 'rrule';
|
||||||
import buildRuleObj, { buildDtStartObj } from './buildRuleObj';
|
import buildRuleObj, { buildDtStartObj } from './buildRuleObj';
|
||||||
|
import { FREQUENCIESCONSTANTS } from './scheduleFormHelpers';
|
||||||
|
|
||||||
window.RRuleSet = RRuleSet;
|
window.RRuleSet = RRuleSet;
|
||||||
|
|
||||||
@@ -12,42 +13,31 @@ export default function buildRuleSet(values, useUTCStart) {
|
|||||||
startDate: values.startDate,
|
startDate: values.startDate,
|
||||||
startTime: values.startTime,
|
startTime: values.startTime,
|
||||||
timezone: values.timezone,
|
timezone: values.timezone,
|
||||||
|
frequency: values.freq,
|
||||||
});
|
});
|
||||||
set.rrule(startRule);
|
set.rrule(startRule);
|
||||||
}
|
}
|
||||||
|
|
||||||
if (values.frequency.length === 0) {
|
values.frequencies.forEach(({ frequency, rrule }) => {
|
||||||
const rule = buildRuleObj(
|
if (!frequencies.includes(frequency)) {
|
||||||
{
|
|
||||||
startDate: values.startDate,
|
|
||||||
startTime: values.startTime,
|
|
||||||
timezone: values.timezone,
|
|
||||||
frequency: 'none',
|
|
||||||
interval: 1,
|
|
||||||
},
|
|
||||||
useUTCStart
|
|
||||||
);
|
|
||||||
set.rrule(new RRule(rule));
|
|
||||||
}
|
|
||||||
|
|
||||||
frequencies.forEach((frequency) => {
|
|
||||||
if (!values.frequency.includes(frequency)) {
|
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
const rule = buildRuleObj(
|
const rule = buildRuleObj(
|
||||||
{
|
{
|
||||||
startDate: values.startDate,
|
startDate: values.startDate,
|
||||||
startTime: values.startTime,
|
startTime: values.startTime,
|
||||||
timezone: values.timezone,
|
timezone: values.timezone,
|
||||||
frequency,
|
freq: FREQUENCIESCONSTANTS[frequency],
|
||||||
...values.frequencyOptions[frequency],
|
rrule,
|
||||||
},
|
},
|
||||||
useUTCStart
|
true
|
||||||
);
|
);
|
||||||
|
|
||||||
set.rrule(new RRule(rule));
|
set.rrule(new RRule(rule));
|
||||||
});
|
});
|
||||||
|
|
||||||
frequencies.forEach((frequency) => {
|
values.exceptions?.forEach(({ frequency, rrule }) => {
|
||||||
if (!values.exceptionFrequency?.includes(frequency)) {
|
if (!values.exceptionFrequency?.includes(frequency)) {
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
@@ -56,8 +46,8 @@ export default function buildRuleSet(values, useUTCStart) {
|
|||||||
startDate: values.startDate,
|
startDate: values.startDate,
|
||||||
startTime: values.startTime,
|
startTime: values.startTime,
|
||||||
timezone: values.timezone,
|
timezone: values.timezone,
|
||||||
frequency,
|
freq: FREQUENCIESCONSTANTS[frequency],
|
||||||
...values.exceptionOptions[frequency],
|
rrule,
|
||||||
},
|
},
|
||||||
useUTCStart
|
useUTCStart
|
||||||
);
|
);
|
||||||
|
|||||||
@@ -12,12 +12,14 @@ export class UnsupportedRRuleError extends Error {
|
|||||||
|
|
||||||
export default function parseRuleObj(schedule) {
|
export default function parseRuleObj(schedule) {
|
||||||
let values = {
|
let values = {
|
||||||
frequency: [],
|
frequency: '',
|
||||||
frequencyOptions: {},
|
rrules: '',
|
||||||
exceptionFrequency: [],
|
|
||||||
exceptionOptions: {},
|
|
||||||
timezone: schedule.timezone,
|
timezone: schedule.timezone,
|
||||||
};
|
};
|
||||||
|
if (Object.values(schedule).length === 0) {
|
||||||
|
return values;
|
||||||
|
}
|
||||||
|
|
||||||
const ruleset = rrulestr(schedule.rrule.replace(' ', '\n'), {
|
const ruleset = rrulestr(schedule.rrule.replace(' ', '\n'), {
|
||||||
forceset: true,
|
forceset: true,
|
||||||
});
|
});
|
||||||
@@ -40,25 +42,9 @@ export default function parseRuleObj(schedule) {
|
|||||||
}
|
}
|
||||||
});
|
});
|
||||||
|
|
||||||
if (isSingleOccurrence(values)) {
|
|
||||||
values.frequency = [];
|
|
||||||
values.frequencyOptions = {};
|
|
||||||
}
|
|
||||||
|
|
||||||
return values;
|
return values;
|
||||||
}
|
}
|
||||||
|
|
||||||
function isSingleOccurrence(values) {
|
|
||||||
if (values.frequency.length > 1) {
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
if (values.frequency[0] !== 'minute') {
|
|
||||||
return false;
|
|
||||||
}
|
|
||||||
const options = values.frequencyOptions.minute;
|
|
||||||
return options.end === 'after' && options.occurrences === 1;
|
|
||||||
}
|
|
||||||
|
|
||||||
function parseDtstart(schedule, values) {
|
function parseDtstart(schedule, values) {
|
||||||
// TODO: should this rely on DTSTART in rruleset rather than schedule.dtstart?
|
// TODO: should this rely on DTSTART in rruleset rather than schedule.dtstart?
|
||||||
const [startDate, startTime] = dateToInputDateTime(
|
const [startDate, startTime] = dateToInputDateTime(
|
||||||
@@ -81,27 +67,12 @@ const frequencyTypes = {
|
|||||||
[RRule.YEARLY]: 'year',
|
[RRule.YEARLY]: 'year',
|
||||||
};
|
};
|
||||||
|
|
||||||
function parseRrule(rruleString, schedule, values) {
|
function parseRrule(rruleString, schedule) {
|
||||||
const { frequency, options } = parseRule(
|
const { frequency } = parseRule(rruleString, schedule);
|
||||||
rruleString,
|
|
||||||
schedule,
|
|
||||||
values.exceptionFrequency
|
|
||||||
);
|
|
||||||
|
|
||||||
if (values.frequencyOptions[frequency]) {
|
const freq = { frequency, rrule: rruleString };
|
||||||
throw new UnsupportedRRuleError(
|
|
||||||
'Duplicate exception frequency types not supported'
|
|
||||||
);
|
|
||||||
}
|
|
||||||
|
|
||||||
return {
|
return freq;
|
||||||
...values,
|
|
||||||
frequency: [...values.frequency, frequency].sort(sortFrequencies),
|
|
||||||
frequencyOptions: {
|
|
||||||
...values.frequencyOptions,
|
|
||||||
[frequency]: options,
|
|
||||||
},
|
|
||||||
};
|
|
||||||
}
|
}
|
||||||
|
|
||||||
function parseExRule(exruleString, schedule, values) {
|
function parseExRule(exruleString, schedule, values) {
|
||||||
@@ -129,20 +100,10 @@ function parseExRule(exruleString, schedule, values) {
|
|||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|
||||||
function parseRule(ruleString, schedule, frequencies) {
|
function parseRule(ruleString, schedule) {
|
||||||
const {
|
const {
|
||||||
origOptions: {
|
origOptions: { count, freq, interval, until, ...rest },
|
||||||
bymonth,
|
|
||||||
bymonthday,
|
|
||||||
bysetpos,
|
|
||||||
byweekday,
|
|
||||||
count,
|
|
||||||
freq,
|
|
||||||
interval,
|
|
||||||
until,
|
|
||||||
},
|
|
||||||
} = RRule.fromString(ruleString);
|
} = RRule.fromString(ruleString);
|
||||||
|
|
||||||
const now = DateTime.now();
|
const now = DateTime.now();
|
||||||
const closestQuarterHour = DateTime.fromMillis(
|
const closestQuarterHour = DateTime.fromMillis(
|
||||||
Math.ceil(now.ts / 900000) * 900000
|
Math.ceil(now.ts / 900000) * 900000
|
||||||
@@ -156,17 +117,17 @@ function parseRule(ruleString, schedule, frequencies) {
|
|||||||
endTime: time,
|
endTime: time,
|
||||||
occurrences: 1,
|
occurrences: 1,
|
||||||
interval: 1,
|
interval: 1,
|
||||||
end: 'never',
|
endingType: 'never',
|
||||||
};
|
};
|
||||||
|
|
||||||
if (until) {
|
if (until?.length) {
|
||||||
options.end = 'onDate';
|
options.endingType = 'onDate';
|
||||||
const end = DateTime.fromISO(until.toISOString());
|
const end = DateTime.fromISO(until.toISOString());
|
||||||
const [endDate, endTime] = dateToInputDateTime(end, schedule.timezone);
|
const [endDate, endTime] = dateToInputDateTime(end, schedule.timezone);
|
||||||
options.endDate = endDate;
|
options.endDate = endDate;
|
||||||
options.endTime = endTime;
|
options.endTime = endTime;
|
||||||
} else if (count) {
|
} else if (count) {
|
||||||
options.end = 'after';
|
options.endingType = 'after';
|
||||||
options.occurrences = count;
|
options.occurrences = count;
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -178,101 +139,10 @@ function parseRule(ruleString, schedule, frequencies) {
|
|||||||
throw new Error(`Unexpected rrule frequency: ${freq}`);
|
throw new Error(`Unexpected rrule frequency: ${freq}`);
|
||||||
}
|
}
|
||||||
const frequency = frequencyTypes[freq];
|
const frequency = frequencyTypes[freq];
|
||||||
if (frequencies.includes(frequency)) {
|
|
||||||
throw new Error(`Duplicate frequency types not supported (${frequency})`);
|
|
||||||
}
|
|
||||||
|
|
||||||
if (freq === RRule.WEEKLY && byweekday) {
|
|
||||||
options.daysOfWeek = byweekday;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (freq === RRule.MONTHLY) {
|
|
||||||
options.runOn = 'day';
|
|
||||||
options.runOnTheOccurrence = 1;
|
|
||||||
options.runOnTheDay = 'sunday';
|
|
||||||
options.runOnDayNumber = 1;
|
|
||||||
|
|
||||||
if (bymonthday) {
|
|
||||||
options.runOnDayNumber = bymonthday;
|
|
||||||
}
|
|
||||||
if (bysetpos) {
|
|
||||||
options.runOn = 'the';
|
|
||||||
options.runOnTheOccurrence = bysetpos;
|
|
||||||
options.runOnTheDay = generateRunOnTheDay(byweekday);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
if (freq === RRule.YEARLY) {
|
|
||||||
options.runOn = 'day';
|
|
||||||
options.runOnTheOccurrence = 1;
|
|
||||||
options.runOnTheDay = 'sunday';
|
|
||||||
options.runOnTheMonth = 1;
|
|
||||||
options.runOnDayMonth = 1;
|
|
||||||
options.runOnDayNumber = 1;
|
|
||||||
|
|
||||||
if (bymonthday) {
|
|
||||||
options.runOnDayNumber = bymonthday;
|
|
||||||
options.runOnDayMonth = bymonth;
|
|
||||||
}
|
|
||||||
if (bysetpos) {
|
|
||||||
options.runOn = 'the';
|
|
||||||
options.runOnTheOccurrence = bysetpos;
|
|
||||||
options.runOnTheDay = generateRunOnTheDay(byweekday);
|
|
||||||
options.runOnTheMonth = bymonth;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return {
|
return {
|
||||||
frequency,
|
frequency,
|
||||||
options,
|
...options,
|
||||||
|
...rest,
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|
||||||
function generateRunOnTheDay(days = []) {
|
|
||||||
if (
|
|
||||||
[
|
|
||||||
RRule.MO,
|
|
||||||
RRule.TU,
|
|
||||||
RRule.WE,
|
|
||||||
RRule.TH,
|
|
||||||
RRule.FR,
|
|
||||||
RRule.SA,
|
|
||||||
RRule.SU,
|
|
||||||
].every((element) => days.indexOf(element) > -1)
|
|
||||||
) {
|
|
||||||
return 'day';
|
|
||||||
}
|
|
||||||
if (
|
|
||||||
[RRule.MO, RRule.TU, RRule.WE, RRule.TH, RRule.FR].every(
|
|
||||||
(element) => days.indexOf(element) > -1
|
|
||||||
)
|
|
||||||
) {
|
|
||||||
return 'weekday';
|
|
||||||
}
|
|
||||||
if ([RRule.SA, RRule.SU].every((element) => days.indexOf(element) > -1)) {
|
|
||||||
return 'weekendDay';
|
|
||||||
}
|
|
||||||
if (days.indexOf(RRule.MO) > -1) {
|
|
||||||
return 'monday';
|
|
||||||
}
|
|
||||||
if (days.indexOf(RRule.TU) > -1) {
|
|
||||||
return 'tuesday';
|
|
||||||
}
|
|
||||||
if (days.indexOf(RRule.WE) > -1) {
|
|
||||||
return 'wednesday';
|
|
||||||
}
|
|
||||||
if (days.indexOf(RRule.TH) > -1) {
|
|
||||||
return 'thursday';
|
|
||||||
}
|
|
||||||
if (days.indexOf(RRule.FR) > -1) {
|
|
||||||
return 'friday';
|
|
||||||
}
|
|
||||||
if (days.indexOf(RRule.SA) > -1) {
|
|
||||||
return 'saturday';
|
|
||||||
}
|
|
||||||
if (days.indexOf(RRule.SU) > -1) {
|
|
||||||
return 'sunday';
|
|
||||||
}
|
|
||||||
|
|
||||||
return null;
|
|
||||||
}
|
|
||||||
|
|||||||
232
awx/ui/src/components/Schedule/shared/scheduleFormHelpers.js
Normal file
232
awx/ui/src/components/Schedule/shared/scheduleFormHelpers.js
Normal file
@@ -0,0 +1,232 @@
|
|||||||
|
import { t } from '@lingui/macro';
|
||||||
|
import { DateTime } from 'luxon';
|
||||||
|
import { RRule } from 'rrule';
|
||||||
|
import buildRuleObj from './buildRuleObj';
|
||||||
|
import buildRuleSet from './buildRuleSet';
|
||||||
|
|
||||||
|
// const NUM_DAYS_PER_FREQUENCY = {
|
||||||
|
// week: 7,
|
||||||
|
// month: 31,
|
||||||
|
// year: 365,
|
||||||
|
// };
|
||||||
|
// const validateSchedule = () =>
|
||||||
|
// const errors = {};
|
||||||
|
|
||||||
|
// values.frequencies.forEach((freq) => {
|
||||||
|
// const options = values.frequencyOptions[freq];
|
||||||
|
// const freqErrors = {};
|
||||||
|
|
||||||
|
// if (
|
||||||
|
// (freq === 'month' || freq === 'year') &&
|
||||||
|
// options.runOn === 'day' &&
|
||||||
|
// (options.runOnDayNumber < 1 || options.runOnDayNumber > 31)
|
||||||
|
// ) {
|
||||||
|
// freqErrors.runOn = t`Please select a day number between 1 and 31.`;
|
||||||
|
// }
|
||||||
|
|
||||||
|
// if (options.end === 'after' && !options.occurrences) {
|
||||||
|
// freqErrors.occurrences = t`Please enter a number of occurrences.`;
|
||||||
|
// }
|
||||||
|
|
||||||
|
// if (options.end === 'onDate') {
|
||||||
|
// if (
|
||||||
|
// DateTime.fromFormat(
|
||||||
|
// `${values.startDate} ${values.startTime}`,
|
||||||
|
// 'yyyy-LL-dd h:mm a'
|
||||||
|
// ).toMillis() >=
|
||||||
|
// DateTime.fromFormat(
|
||||||
|
// `${options.endDate} ${options.endTime}`,
|
||||||
|
// 'yyyy-LL-dd h:mm a'
|
||||||
|
// ).toMillis()
|
||||||
|
// ) {
|
||||||
|
// freqErrors.endDate = t`Please select an end date/time that comes after the start date/time.`;
|
||||||
|
// }
|
||||||
|
|
||||||
|
// if (
|
||||||
|
// DateTime.fromISO(options.endDate)
|
||||||
|
// .diff(DateTime.fromISO(values.startDate), 'days')
|
||||||
|
// .toObject().days < NUM_DAYS_PER_FREQUENCY[freq]
|
||||||
|
// ) {
|
||||||
|
// const rule = new RRule(
|
||||||
|
// buildRuleObj({
|
||||||
|
// startDate: values.startDate,
|
||||||
|
// startTime: values.startTime,
|
||||||
|
// frequencies: freq,
|
||||||
|
// ...options,
|
||||||
|
// })
|
||||||
|
// );
|
||||||
|
// if (rule.all().length === 0) {
|
||||||
|
// errors.startDate = t`Selected date range must have at least 1 schedule occurrence.`;
|
||||||
|
// freqErrors.endDate = t`Selected date range must have at least 1 schedule occurrence.`;
|
||||||
|
// }
|
||||||
|
// }
|
||||||
|
// }
|
||||||
|
// if (Object.keys(freqErrors).length > 0) {
|
||||||
|
// if (!errors.frequencyOptions) {
|
||||||
|
// errors.frequencyOptions = {};
|
||||||
|
// }
|
||||||
|
// errors.frequencyOptions[freq] = freqErrors;
|
||||||
|
// }
|
||||||
|
// });
|
||||||
|
|
||||||
|
// if (values.exceptionFrequency.length > 0 && !scheduleHasInstances(values)) {
|
||||||
|
// errors.exceptionFrequency = t`This schedule has no occurrences due to the
|
||||||
|
// selected exceptions.`;
|
||||||
|
// }
|
||||||
|
|
||||||
|
// ({});
|
||||||
|
// function scheduleHasInstances(values) {
|
||||||
|
// let rangeToCheck = 1;
|
||||||
|
// values.frequencies.forEach((freq) => {
|
||||||
|
// if (NUM_DAYS_PER_FREQUENCY[freq] > rangeToCheck) {
|
||||||
|
// rangeToCheck = NUM_DAYS_PER_FREQUENCY[freq];
|
||||||
|
// }
|
||||||
|
// });
|
||||||
|
|
||||||
|
// const ruleSet = buildRuleSet(values, true);
|
||||||
|
// const startDate = DateTime.fromISO(values.startDate);
|
||||||
|
// const endDate = startDate.plus({ days: rangeToCheck });
|
||||||
|
// const instances = ruleSet.between(
|
||||||
|
// startDate.toJSDate(),
|
||||||
|
// endDate.toJSDate(),
|
||||||
|
// true,
|
||||||
|
// (date, i) => i === 0
|
||||||
|
// );
|
||||||
|
|
||||||
|
// return instances.length > 0;
|
||||||
|
// }
|
||||||
|
|
||||||
|
const bysetposOptions = [
|
||||||
|
{ value: '', key: 'none', label: 'None' },
|
||||||
|
{ value: 1, key: 'first', label: t`First` },
|
||||||
|
{
|
||||||
|
value: 2,
|
||||||
|
key: 'second',
|
||||||
|
label: t`Second`,
|
||||||
|
},
|
||||||
|
{ value: 3, key: 'third', label: t`Third` },
|
||||||
|
{
|
||||||
|
value: 4,
|
||||||
|
key: 'fourth',
|
||||||
|
label: t`Fourth`,
|
||||||
|
},
|
||||||
|
{ value: 5, key: 'fifth', label: t`Fifth` },
|
||||||
|
{ value: -1, key: 'last', label: t`Last` },
|
||||||
|
];
|
||||||
|
|
||||||
|
const monthOptions = [
|
||||||
|
{
|
||||||
|
key: 'january',
|
||||||
|
value: 1,
|
||||||
|
label: t`January`,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
key: 'february',
|
||||||
|
value: 2,
|
||||||
|
label: t`February`,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
key: 'march',
|
||||||
|
value: 3,
|
||||||
|
label: t`March`,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
key: 'april',
|
||||||
|
value: 4,
|
||||||
|
label: t`April`,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
key: 'may',
|
||||||
|
value: 5,
|
||||||
|
label: t`May`,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
key: 'june',
|
||||||
|
value: 6,
|
||||||
|
label: t`June`,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
key: 'july',
|
||||||
|
value: 7,
|
||||||
|
label: t`July`,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
key: 'august',
|
||||||
|
value: 8,
|
||||||
|
label: t`August`,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
key: 'september',
|
||||||
|
value: 9,
|
||||||
|
label: t`September`,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
key: 'october',
|
||||||
|
value: 10,
|
||||||
|
label: t`October`,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
key: 'november',
|
||||||
|
value: 11,
|
||||||
|
label: t`November`,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
key: 'december',
|
||||||
|
value: 12,
|
||||||
|
label: t`December`,
|
||||||
|
},
|
||||||
|
];
|
||||||
|
|
||||||
|
const weekdayOptions = [
|
||||||
|
{
|
||||||
|
value: RRule.SU,
|
||||||
|
key: 'sunday',
|
||||||
|
label: t`Sunday`,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
value: RRule.MO,
|
||||||
|
key: 'monday',
|
||||||
|
label: t`Monday`,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
value: RRule.TU,
|
||||||
|
key: 'tuesday',
|
||||||
|
label: t`Tuesday`,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
value: RRule.WE,
|
||||||
|
key: 'wednesday',
|
||||||
|
label: t`Wednesday`,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
value: RRule.TH,
|
||||||
|
key: 'thursday',
|
||||||
|
label: t`Thursday`,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
value: RRule.FR,
|
||||||
|
key: 'friday',
|
||||||
|
label: t`Friday`,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
value: RRule.SA,
|
||||||
|
key: 'saturday',
|
||||||
|
label: t`Saturday`,
|
||||||
|
},
|
||||||
|
];
|
||||||
|
|
||||||
|
const FREQUENCIESCONSTANTS = {
|
||||||
|
minute: RRule.MINUTELY,
|
||||||
|
hour: RRule.HOURLY,
|
||||||
|
day: RRule.DAILY,
|
||||||
|
week: RRule.WEEKLY,
|
||||||
|
month: RRule.MONTHLY,
|
||||||
|
year: RRule.YEARLY,
|
||||||
|
};
|
||||||
|
export {
|
||||||
|
monthOptions,
|
||||||
|
weekdayOptions,
|
||||||
|
bysetposOptions,
|
||||||
|
// validateSchedule,
|
||||||
|
FREQUENCIESCONSTANTS,
|
||||||
|
};
|
||||||
@@ -24,12 +24,10 @@ function WorkflowOutputNavigation({ relatedJobs, parentRef }) {
|
|||||||
const { id } = useParams();
|
const { id } = useParams();
|
||||||
|
|
||||||
const relevantResults = relatedJobs.filter(
|
const relevantResults = relatedJobs.filter(
|
||||||
({
|
({ job: jobId, summary_fields }) =>
|
||||||
job: jobId,
|
jobId &&
|
||||||
summary_fields: {
|
`${jobId}` !== id &&
|
||||||
unified_job_template: { unified_job_type },
|
summary_fields.job.type !== 'workflow_approval'
|
||||||
},
|
|
||||||
}) => jobId && `${jobId}` !== id && unified_job_type !== 'workflow_approval'
|
|
||||||
);
|
);
|
||||||
|
|
||||||
const [isOpen, setIsOpen] = useState(false);
|
const [isOpen, setIsOpen] = useState(false);
|
||||||
@@ -101,16 +99,14 @@ function WorkflowOutputNavigation({ relatedJobs, parentRef }) {
|
|||||||
{sortedJobs?.map((node) => (
|
{sortedJobs?.map((node) => (
|
||||||
<SelectOption
|
<SelectOption
|
||||||
key={node.id}
|
key={node.id}
|
||||||
to={`/jobs/${
|
to={`/jobs/${JOB_URL_SEGMENT_MAP[node.summary_fields.job.type]}/${
|
||||||
JOB_URL_SEGMENT_MAP[
|
node.summary_fields.job?.id
|
||||||
node.summary_fields.unified_job_template.unified_job_type
|
}/output`}
|
||||||
]
|
|
||||||
}/${node.summary_fields.job?.id}/output`}
|
|
||||||
component={Link}
|
component={Link}
|
||||||
value={node.summary_fields.unified_job_template.name}
|
value={node.summary_fields.job.name}
|
||||||
>
|
>
|
||||||
{stringIsUUID(node.identifier)
|
{stringIsUUID(node.identifier)
|
||||||
? node.summary_fields.unified_job_template.name
|
? node.summary_fields.job.name
|
||||||
: node.identifier}
|
: node.identifier}
|
||||||
</SelectOption>
|
</SelectOption>
|
||||||
))}
|
))}
|
||||||
|
|||||||
@@ -0,0 +1,85 @@
|
|||||||
|
import React from 'react';
|
||||||
|
import { within, render, screen, waitFor } from '@testing-library/react';
|
||||||
|
import userEvent from '@testing-library/user-event';
|
||||||
|
import WorkflowOutputNavigation from './WorkflowOutputNavigation';
|
||||||
|
import { createMemoryHistory } from 'history';
|
||||||
|
import { I18nProvider } from '@lingui/react';
|
||||||
|
import { i18n } from '@lingui/core';
|
||||||
|
import { en } from 'make-plural/plurals';
|
||||||
|
import english from '../../../src/locales/en/messages';
|
||||||
|
import { Router } from 'react-router-dom';
|
||||||
|
|
||||||
|
jest.mock('react-router-dom', () => ({
|
||||||
|
...jest.requireActual('react-router-dom'),
|
||||||
|
useParams: () => ({
|
||||||
|
id: 1,
|
||||||
|
}),
|
||||||
|
}));
|
||||||
|
const jobs = [
|
||||||
|
{
|
||||||
|
id: 1,
|
||||||
|
summary_fields: {
|
||||||
|
job: {
|
||||||
|
name: 'Ansible',
|
||||||
|
type: 'project_update',
|
||||||
|
id: 1,
|
||||||
|
status: 'successful',
|
||||||
|
},
|
||||||
|
},
|
||||||
|
job: 4,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
id: 2,
|
||||||
|
summary_fields: {
|
||||||
|
job: {
|
||||||
|
name: 'Durham',
|
||||||
|
type: 'job',
|
||||||
|
id: 2,
|
||||||
|
status: 'successful',
|
||||||
|
},
|
||||||
|
},
|
||||||
|
job: 3,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
id: 3,
|
||||||
|
summary_fields: {
|
||||||
|
job: {
|
||||||
|
name: 'Red hat',
|
||||||
|
type: 'job',
|
||||||
|
id: 3,
|
||||||
|
status: 'successful',
|
||||||
|
},
|
||||||
|
},
|
||||||
|
job: 2,
|
||||||
|
},
|
||||||
|
];
|
||||||
|
|
||||||
|
describe('<WorkflowOuputNavigation/>', () => {
|
||||||
|
test('Should open modal and deprovision node', async () => {
|
||||||
|
i18n.loadLocaleData({ en: { plurals: en } });
|
||||||
|
i18n.load({ en: english });
|
||||||
|
i18n.activate('en');
|
||||||
|
const user = userEvent.setup();
|
||||||
|
const ref = jest
|
||||||
|
.spyOn(React, 'useRef')
|
||||||
|
.mockReturnValueOnce({ current: 'div' });
|
||||||
|
const history = createMemoryHistory({
|
||||||
|
initialEntries: ['jobs/playbook/2/output'],
|
||||||
|
});
|
||||||
|
render(
|
||||||
|
<I18nProvider i18n={i18n}>
|
||||||
|
<Router history={history}>
|
||||||
|
<WorkflowOutputNavigation relatedJobs={jobs} parentRef={ref} />
|
||||||
|
</Router>
|
||||||
|
</I18nProvider>
|
||||||
|
);
|
||||||
|
|
||||||
|
const button = screen.getByRole('button');
|
||||||
|
await user.click(button);
|
||||||
|
|
||||||
|
await waitFor(() => screen.getByText('Workflow Nodes'));
|
||||||
|
await waitFor(() => screen.getByText('Red hat'));
|
||||||
|
await waitFor(() => screen.getByText('Durham'));
|
||||||
|
await waitFor(() => screen.getByText('Ansible'));
|
||||||
|
});
|
||||||
|
});
|
||||||
@@ -465,7 +465,7 @@
|
|||||||
},
|
},
|
||||||
"created": "2020-05-18T21:53:35.370730Z",
|
"created": "2020-05-18T21:53:35.370730Z",
|
||||||
"modified": "2020-05-18T21:54:05.436400Z",
|
"modified": "2020-05-18T21:54:05.436400Z",
|
||||||
"name": "CyberArk AIM Central Credential Provider Lookup",
|
"name": "CyberArk Central Credential Provider Lookup",
|
||||||
"description": "",
|
"description": "",
|
||||||
"kind": "external",
|
"kind": "external",
|
||||||
"namespace": "aim",
|
"namespace": "aim",
|
||||||
|
|||||||
@@ -29,6 +29,10 @@ function ContainerGroupAdd() {
|
|||||||
try {
|
try {
|
||||||
const { data: response } = await InstanceGroupsAPI.create({
|
const { data: response } = await InstanceGroupsAPI.create({
|
||||||
name: values.name,
|
name: values.name,
|
||||||
|
max_forks: values.max_forks ? values.max_forks : 0,
|
||||||
|
max_concurrent_jobs: values.max_concurrent_jobs
|
||||||
|
? values.max_concurrent_jobs
|
||||||
|
: 0,
|
||||||
credential: values?.credential?.id,
|
credential: values?.credential?.id,
|
||||||
pod_spec_override: values.override
|
pod_spec_override: values.override
|
||||||
? getPodSpecValue(values.pod_spec_override)
|
? getPodSpecValue(values.pod_spec_override)
|
||||||
|
|||||||
@@ -33,6 +33,8 @@ const initialPodSpec = {
|
|||||||
const instanceGroupCreateData = {
|
const instanceGroupCreateData = {
|
||||||
name: 'Fuz',
|
name: 'Fuz',
|
||||||
credential: { id: 71, name: 'CG' },
|
credential: { id: 71, name: 'CG' },
|
||||||
|
max_concurrent_jobs: 0,
|
||||||
|
max_forks: 0,
|
||||||
pod_spec_override:
|
pod_spec_override:
|
||||||
'apiVersion: v1\nkind: Pod\nmetadata:\n namespace: default\nspec:\n containers:\n - image: ansible/ansible-runner\n tty: true\n stdin: true\n imagePullPolicy: Always\n args:\n - sleep\n - infinity\n - test',
|
'apiVersion: v1\nkind: Pod\nmetadata:\n namespace: default\nspec:\n containers:\n - image: ansible/ansible-runner\n tty: true\n stdin: true\n imagePullPolicy: Always\n args:\n - sleep\n - infinity\n - test',
|
||||||
};
|
};
|
||||||
|
|||||||
@@ -9,7 +9,12 @@ import AlertModal from 'components/AlertModal';
|
|||||||
import ErrorDetail from 'components/ErrorDetail';
|
import ErrorDetail from 'components/ErrorDetail';
|
||||||
import { CardBody, CardActionsRow } from 'components/Card';
|
import { CardBody, CardActionsRow } from 'components/Card';
|
||||||
import DeleteButton from 'components/DeleteButton';
|
import DeleteButton from 'components/DeleteButton';
|
||||||
import { Detail, DetailList, UserDateDetail } from 'components/DetailList';
|
import {
|
||||||
|
Detail,
|
||||||
|
DetailList,
|
||||||
|
UserDateDetail,
|
||||||
|
DetailBadge,
|
||||||
|
} from 'components/DetailList';
|
||||||
import useRequest, { useDismissableError } from 'hooks/useRequest';
|
import useRequest, { useDismissableError } from 'hooks/useRequest';
|
||||||
import { jsonToYaml, isJsonString } from 'util/yaml';
|
import { jsonToYaml, isJsonString } from 'util/yaml';
|
||||||
import { InstanceGroupsAPI } from 'api';
|
import { InstanceGroupsAPI } from 'api';
|
||||||
@@ -47,6 +52,20 @@ function ContainerGroupDetails({ instanceGroup }) {
|
|||||||
value={t`Container group`}
|
value={t`Container group`}
|
||||||
dataCy="container-group-type"
|
dataCy="container-group-type"
|
||||||
/>
|
/>
|
||||||
|
<DetailBadge
|
||||||
|
label={t`Max concurrent jobs`}
|
||||||
|
dataCy="instance-group-max-concurrent-jobs"
|
||||||
|
helpText={t`Maximum number of jobs to run concurrently on this group.
|
||||||
|
Zero means no limit will be enforced.`}
|
||||||
|
content={instanceGroup.max_concurrent_jobs}
|
||||||
|
/>
|
||||||
|
<DetailBadge
|
||||||
|
label={t`Max forks`}
|
||||||
|
dataCy="instance-group-max-forks"
|
||||||
|
helpText={t`Maximum number of forks to allow across all jobs running concurrently on this group.
|
||||||
|
Zero means no limit will be enforced.`}
|
||||||
|
content={instanceGroup.max_forks}
|
||||||
|
/>
|
||||||
{instanceGroup.summary_fields.credential && (
|
{instanceGroup.summary_fields.credential && (
|
||||||
<Detail
|
<Detail
|
||||||
label={t`Credential`}
|
label={t`Credential`}
|
||||||
|
|||||||
@@ -23,6 +23,8 @@ const instanceGroup = {
|
|||||||
created: '2020-09-03T18:26:47.113934Z',
|
created: '2020-09-03T18:26:47.113934Z',
|
||||||
modified: '2020-09-03T19:34:23.244694Z',
|
modified: '2020-09-03T19:34:23.244694Z',
|
||||||
capacity: 0,
|
capacity: 0,
|
||||||
|
max_concurrent_jobs: 0,
|
||||||
|
max_forks: 0,
|
||||||
committed_capacity: 0,
|
committed_capacity: 0,
|
||||||
consumed_capacity: 0,
|
consumed_capacity: 0,
|
||||||
percent_capacity_remaining: 0.0,
|
percent_capacity_remaining: 0.0,
|
||||||
|
|||||||
@@ -39,6 +39,10 @@ function ContainerGroupEdit({ instanceGroup }) {
|
|||||||
name: values.name,
|
name: values.name,
|
||||||
credential: values.credential ? values.credential.id : null,
|
credential: values.credential ? values.credential.id : null,
|
||||||
pod_spec_override: values.override ? values.pod_spec_override : null,
|
pod_spec_override: values.override ? values.pod_spec_override : null,
|
||||||
|
max_forks: values.max_forks ? values.max_forks : 0,
|
||||||
|
max_concurrent_jobs: values.max_concurrent_jobs
|
||||||
|
? values.max_concurrent_jobs
|
||||||
|
: 0,
|
||||||
is_container_group: true,
|
is_container_group: true,
|
||||||
});
|
});
|
||||||
history.push(detailsIUrl);
|
history.push(detailsIUrl);
|
||||||
|
|||||||
@@ -34,6 +34,8 @@ const instanceGroup = {
|
|||||||
policy_instance_percentage: 0,
|
policy_instance_percentage: 0,
|
||||||
policy_instance_minimum: 0,
|
policy_instance_minimum: 0,
|
||||||
policy_instance_list: [],
|
policy_instance_list: [],
|
||||||
|
max_concurrent_jobs: 0,
|
||||||
|
max_forks: 0,
|
||||||
pod_spec_override: '',
|
pod_spec_override: '',
|
||||||
summary_fields: {
|
summary_fields: {
|
||||||
credential: {
|
credential: {
|
||||||
@@ -144,6 +146,8 @@ describe('<ContainerGroupEdit/>', () => {
|
|||||||
...updatedInstanceGroup,
|
...updatedInstanceGroup,
|
||||||
credential: 12,
|
credential: 12,
|
||||||
pod_spec_override: null,
|
pod_spec_override: null,
|
||||||
|
max_concurrent_jobs: 0,
|
||||||
|
max_forks: 0,
|
||||||
is_container_group: true,
|
is_container_group: true,
|
||||||
});
|
});
|
||||||
expect(history.location.pathname).toEqual(
|
expect(history.location.pathname).toEqual(
|
||||||
|
|||||||
@@ -42,6 +42,8 @@ const instanceGroup = {
|
|||||||
credential: null,
|
credential: null,
|
||||||
policy_instance_percentage: 100,
|
policy_instance_percentage: 100,
|
||||||
policy_instance_minimum: 0,
|
policy_instance_minimum: 0,
|
||||||
|
max_concurrent_jobs: 0,
|
||||||
|
max_forks: 0,
|
||||||
policy_instance_list: ['receptor-1', 'receptor-2'],
|
policy_instance_list: ['receptor-1', 'receptor-2'],
|
||||||
pod_spec_override: '',
|
pod_spec_override: '',
|
||||||
summary_fields: {
|
summary_fields: {
|
||||||
|
|||||||
@@ -73,6 +73,20 @@ function InstanceGroupDetails({ instanceGroup }) {
|
|||||||
dataCy="instance-group-policy-instance-percentage"
|
dataCy="instance-group-policy-instance-percentage"
|
||||||
content={`${instanceGroup.policy_instance_percentage} %`}
|
content={`${instanceGroup.policy_instance_percentage} %`}
|
||||||
/>
|
/>
|
||||||
|
<DetailBadge
|
||||||
|
label={t`Max concurrent jobs`}
|
||||||
|
dataCy="instance-group-max-concurrent-jobs"
|
||||||
|
helpText={t`Maximum number of jobs to run concurrently on this group.
|
||||||
|
Zero means no limit will be enforced.`}
|
||||||
|
content={instanceGroup.max_concurrent_jobs}
|
||||||
|
/>
|
||||||
|
<DetailBadge
|
||||||
|
label={t`Max forks`}
|
||||||
|
dataCy="instance-group-max-forks"
|
||||||
|
helpText={t`Maximum number of forks to allow across all jobs running concurrently on this group.
|
||||||
|
Zero means no limit will be enforced.`}
|
||||||
|
content={instanceGroup.max_forks}
|
||||||
|
/>
|
||||||
{instanceGroup.capacity ? (
|
{instanceGroup.capacity ? (
|
||||||
<DetailBadge
|
<DetailBadge
|
||||||
label={t`Used capacity`}
|
label={t`Used capacity`}
|
||||||
|
|||||||
@@ -19,6 +19,8 @@ const instanceGroups = [
|
|||||||
policy_instance_minimum: 10,
|
policy_instance_minimum: 10,
|
||||||
policy_instance_percentage: 50,
|
policy_instance_percentage: 50,
|
||||||
percent_capacity_remaining: 60,
|
percent_capacity_remaining: 60,
|
||||||
|
max_concurrent_jobs: 0,
|
||||||
|
max_forks: 0,
|
||||||
is_container_group: false,
|
is_container_group: false,
|
||||||
created: '2020-07-21T18:41:02.818081Z',
|
created: '2020-07-21T18:41:02.818081Z',
|
||||||
modified: '2020-07-24T20:32:03.121079Z',
|
modified: '2020-07-24T20:32:03.121079Z',
|
||||||
@@ -38,6 +40,8 @@ const instanceGroups = [
|
|||||||
policy_instance_minimum: 0,
|
policy_instance_minimum: 0,
|
||||||
policy_instance_percentage: 0,
|
policy_instance_percentage: 0,
|
||||||
percent_capacity_remaining: 0,
|
percent_capacity_remaining: 0,
|
||||||
|
max_concurrent_jobs: 0,
|
||||||
|
max_forks: 0,
|
||||||
is_container_group: true,
|
is_container_group: true,
|
||||||
created: '2020-07-21T18:41:02.818081Z',
|
created: '2020-07-21T18:41:02.818081Z',
|
||||||
modified: '2020-07-24T20:32:03.121079Z',
|
modified: '2020-07-24T20:32:03.121079Z',
|
||||||
|
|||||||
@@ -11,7 +11,7 @@ import FormField, {
|
|||||||
CheckboxField,
|
CheckboxField,
|
||||||
} from 'components/FormField';
|
} from 'components/FormField';
|
||||||
import FormActionGroup from 'components/FormActionGroup';
|
import FormActionGroup from 'components/FormActionGroup';
|
||||||
import { required } from 'util/validators';
|
import { required, minMaxValue } from 'util/validators';
|
||||||
import {
|
import {
|
||||||
FormColumnLayout,
|
FormColumnLayout,
|
||||||
FormFullWidthLayout,
|
FormFullWidthLayout,
|
||||||
@@ -57,6 +57,26 @@ function ContainerGroupFormFields({ instanceGroup }) {
|
|||||||
tooltip={t`Credential to authenticate with Kubernetes or OpenShift. Must be of type "Kubernetes/OpenShift API Bearer Token". If left blank, the underlying Pod's service account will be used.`}
|
tooltip={t`Credential to authenticate with Kubernetes or OpenShift. Must be of type "Kubernetes/OpenShift API Bearer Token". If left blank, the underlying Pod's service account will be used.`}
|
||||||
autoPopulate={!instanceGroup?.id}
|
autoPopulate={!instanceGroup?.id}
|
||||||
/>
|
/>
|
||||||
|
<FormField
|
||||||
|
id="instance-group-max-concurrent-jobs"
|
||||||
|
label={t`Max concurrent jobs`}
|
||||||
|
name="max_concurrent_jobs"
|
||||||
|
type="number"
|
||||||
|
min="0"
|
||||||
|
validate={minMaxValue(0, 2147483647)}
|
||||||
|
tooltip={t`Maximum number of jobs to run concurrently on this group.
|
||||||
|
Zero means no limit will be enforced.`}
|
||||||
|
/>
|
||||||
|
<FormField
|
||||||
|
id="instance-group-max-forks"
|
||||||
|
label={t`Max forks`}
|
||||||
|
name="max_forks"
|
||||||
|
type="number"
|
||||||
|
min="0"
|
||||||
|
validate={minMaxValue(0, 2147483647)}
|
||||||
|
tooltip={t`Maximum number of forks to allow across all jobs running concurrently on this group.
|
||||||
|
Zero means no limit will be enforced.`}
|
||||||
|
/>
|
||||||
|
|
||||||
<FormGroup fieldId="container-groups-option-checkbox" label={t`Options`}>
|
<FormGroup fieldId="container-groups-option-checkbox" label={t`Options`}>
|
||||||
<FormCheckboxLayout>
|
<FormCheckboxLayout>
|
||||||
@@ -97,6 +117,8 @@ function ContainerGroupForm({
|
|||||||
|
|
||||||
const initialValues = {
|
const initialValues = {
|
||||||
name: instanceGroup?.name || '',
|
name: instanceGroup?.name || '',
|
||||||
|
max_concurrent_jobs: instanceGroup.max_concurrent_jobs || 0,
|
||||||
|
max_forks: instanceGroup.max_forks || 0,
|
||||||
credential: instanceGroup?.summary_fields?.credential,
|
credential: instanceGroup?.summary_fields?.credential,
|
||||||
pod_spec_override: isCheckboxChecked
|
pod_spec_override: isCheckboxChecked
|
||||||
? instanceGroup?.pod_spec_override
|
? instanceGroup?.pod_spec_override
|
||||||
|
|||||||
@@ -42,6 +42,26 @@ function InstanceGroupFormFields() {
|
|||||||
assigned to this group when new instances come online.`}
|
assigned to this group when new instances come online.`}
|
||||||
validate={minMaxValue(0, 100)}
|
validate={minMaxValue(0, 100)}
|
||||||
/>
|
/>
|
||||||
|
<FormField
|
||||||
|
id="instance-group-max-concurrent-jobs"
|
||||||
|
label={t`Max concurrent jobs`}
|
||||||
|
name="max_concurrent_jobs"
|
||||||
|
type="number"
|
||||||
|
min="0"
|
||||||
|
validate={minMaxValue(0, 2147483647)}
|
||||||
|
tooltip={t`Maximum number of jobs to run concurrently on this group.
|
||||||
|
Zero means no limit will be enforced.`}
|
||||||
|
/>
|
||||||
|
<FormField
|
||||||
|
id="instance-group-max-forks"
|
||||||
|
label={t`Max forks`}
|
||||||
|
name="max_forks"
|
||||||
|
type="number"
|
||||||
|
min="0"
|
||||||
|
validate={minMaxValue(0, 2147483647)}
|
||||||
|
tooltip={t`Maximum number of forks to allow across all jobs running concurrently on this group.
|
||||||
|
Zero means no limit will be enforced.`}
|
||||||
|
/>
|
||||||
</>
|
</>
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
@@ -57,6 +77,8 @@ function InstanceGroupForm({
|
|||||||
name: instanceGroup.name || '',
|
name: instanceGroup.name || '',
|
||||||
policy_instance_minimum: instanceGroup.policy_instance_minimum || 0,
|
policy_instance_minimum: instanceGroup.policy_instance_minimum || 0,
|
||||||
policy_instance_percentage: instanceGroup.policy_instance_percentage || 0,
|
policy_instance_percentage: instanceGroup.policy_instance_percentage || 0,
|
||||||
|
max_concurrent_jobs: instanceGroup.max_concurrent_jobs || 0,
|
||||||
|
max_forks: instanceGroup.max_forks || 0,
|
||||||
};
|
};
|
||||||
return (
|
return (
|
||||||
<Formik
|
<Formik
|
||||||
|
|||||||
@@ -1,13 +1,17 @@
|
|||||||
import React from 'react';
|
import React, { useCallback } from 'react';
|
||||||
import { string, bool, func } from 'prop-types';
|
import { string, bool, func } from 'prop-types';
|
||||||
import { t } from '@lingui/macro';
|
import { t } from '@lingui/macro';
|
||||||
import { Button } from '@patternfly/react-core';
|
|
||||||
import { Tr, Td } from '@patternfly/react-table';
|
import { Tr, Td } from '@patternfly/react-table';
|
||||||
import { Link } from 'react-router-dom';
|
import { Link } from 'react-router-dom';
|
||||||
import { PencilAltIcon } from '@patternfly/react-icons';
|
import { PencilAltIcon } from '@patternfly/react-icons';
|
||||||
import { ActionsTd, ActionItem, TdBreakWord } from 'components/PaginatedTable';
|
import { Button, Chip } from '@patternfly/react-core';
|
||||||
|
import { HostsAPI } from 'api';
|
||||||
|
import AlertModal from 'components/AlertModal';
|
||||||
|
import ChipGroup from 'components/ChipGroup';
|
||||||
|
import ErrorDetail from 'components/ErrorDetail';
|
||||||
import HostToggle from 'components/HostToggle';
|
import HostToggle from 'components/HostToggle';
|
||||||
|
import { ActionsTd, ActionItem, TdBreakWord } from 'components/PaginatedTable';
|
||||||
|
import useRequest, { useDismissableError } from 'hooks/useRequest';
|
||||||
import { Host } from 'types';
|
import { Host } from 'types';
|
||||||
|
|
||||||
function InventoryHostItem({
|
function InventoryHostItem({
|
||||||
@@ -19,8 +23,34 @@ function InventoryHostItem({
|
|||||||
rowIndex,
|
rowIndex,
|
||||||
}) {
|
}) {
|
||||||
const labelId = `check-action-${host.id}`;
|
const labelId = `check-action-${host.id}`;
|
||||||
|
const initialGroups = host?.summary_fields?.groups ?? {
|
||||||
|
results: [],
|
||||||
|
count: 0,
|
||||||
|
};
|
||||||
|
|
||||||
|
const {
|
||||||
|
error,
|
||||||
|
request: fetchRelatedGroups,
|
||||||
|
result: relatedGroups,
|
||||||
|
} = useRequest(
|
||||||
|
useCallback(async (hostId) => {
|
||||||
|
const { data } = await HostsAPI.readGroups(hostId);
|
||||||
|
return data.results;
|
||||||
|
}, []),
|
||||||
|
initialGroups.results
|
||||||
|
);
|
||||||
|
|
||||||
|
const { error: dismissableError, dismissError } = useDismissableError(error);
|
||||||
|
|
||||||
|
const handleOverflowChipClick = (hostId) => {
|
||||||
|
if (relatedGroups.length === initialGroups.count) {
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
fetchRelatedGroups(hostId);
|
||||||
|
};
|
||||||
|
|
||||||
return (
|
return (
|
||||||
|
<>
|
||||||
<Tr id={`host-row-${host.id}`} ouiaId={`inventory-host-row-${host.id}`}>
|
<Tr id={`host-row-${host.id}`} ouiaId={`inventory-host-row-${host.id}`}>
|
||||||
<Td
|
<Td
|
||||||
data-cy={labelId}
|
data-cy={labelId}
|
||||||
@@ -41,13 +71,36 @@ function InventoryHostItem({
|
|||||||
>
|
>
|
||||||
{host.description}
|
{host.description}
|
||||||
</TdBreakWord>
|
</TdBreakWord>
|
||||||
<ActionsTd dataLabel={t`Actions`} gridColumns="auto 40px">
|
<TdBreakWord
|
||||||
|
id={`host-related-groups-${host.id}`}
|
||||||
|
dataLabel={t`Related Groups`}
|
||||||
|
>
|
||||||
|
<ChipGroup
|
||||||
|
aria-label={t`Related Groups`}
|
||||||
|
numChips={4}
|
||||||
|
totalChips={initialGroups.count}
|
||||||
|
ouiaId="host-related-groups-chips"
|
||||||
|
onOverflowChipClick={() => handleOverflowChipClick(host.id)}
|
||||||
|
>
|
||||||
|
{relatedGroups.map((group) => (
|
||||||
|
<Chip key={group.name} isReadOnly>
|
||||||
|
{group.name}
|
||||||
|
</Chip>
|
||||||
|
))}
|
||||||
|
</ChipGroup>
|
||||||
|
</TdBreakWord>
|
||||||
|
<ActionsTd
|
||||||
|
aria-label={t`Actions`}
|
||||||
|
dataLabel={t`Actions`}
|
||||||
|
gridColumns="auto 40px"
|
||||||
|
>
|
||||||
<HostToggle host={host} />
|
<HostToggle host={host} />
|
||||||
<ActionItem
|
<ActionItem
|
||||||
visible={host.summary_fields.user_capabilities?.edit}
|
visible={host.summary_fields.user_capabilities?.edit}
|
||||||
tooltip={t`Edit host`}
|
tooltip={t`Edit host`}
|
||||||
>
|
>
|
||||||
<Button
|
<Button
|
||||||
|
aria-label={t`Edit host`}
|
||||||
ouiaId={`${host.id}-edit-button`}
|
ouiaId={`${host.id}-edit-button`}
|
||||||
variant="plain"
|
variant="plain"
|
||||||
component={Link}
|
component={Link}
|
||||||
@@ -58,6 +111,18 @@ function InventoryHostItem({
|
|||||||
</ActionItem>
|
</ActionItem>
|
||||||
</ActionsTd>
|
</ActionsTd>
|
||||||
</Tr>
|
</Tr>
|
||||||
|
{dismissableError && (
|
||||||
|
<AlertModal
|
||||||
|
isOpen={dismissableError}
|
||||||
|
onClose={dismissError}
|
||||||
|
title={t`Error!`}
|
||||||
|
variant="error"
|
||||||
|
>
|
||||||
|
{t`Failed to load related groups.`}
|
||||||
|
<ErrorDetail error={dismissableError} />
|
||||||
|
</AlertModal>
|
||||||
|
)}
|
||||||
|
</>
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -1,6 +1,21 @@
|
|||||||
import React from 'react';
|
import React from 'react';
|
||||||
import { mountWithContexts } from '../../../../testUtils/enzymeHelpers';
|
import { Router } from 'react-router-dom';
|
||||||
|
import {
|
||||||
|
render,
|
||||||
|
fireEvent,
|
||||||
|
screen,
|
||||||
|
waitFor,
|
||||||
|
within,
|
||||||
|
} from '@testing-library/react';
|
||||||
|
import '@testing-library/jest-dom';
|
||||||
|
import { HostsAPI } from 'api';
|
||||||
|
import { i18n } from '@lingui/core';
|
||||||
|
import { en } from 'make-plural/plurals';
|
||||||
import InventoryHostItem from './InventoryHostItem';
|
import InventoryHostItem from './InventoryHostItem';
|
||||||
|
import { createMemoryHistory } from 'history';
|
||||||
|
import english from '../../../locales/en/messages';
|
||||||
|
|
||||||
|
jest.mock('api');
|
||||||
|
|
||||||
const mockHost = {
|
const mockHost = {
|
||||||
id: 1,
|
id: 1,
|
||||||
@@ -24,58 +39,194 @@ const mockHost = {
|
|||||||
finished: '2020-02-26T22:38:41.037991Z',
|
finished: '2020-02-26T22:38:41.037991Z',
|
||||||
},
|
},
|
||||||
],
|
],
|
||||||
|
groups: {
|
||||||
|
count: 1,
|
||||||
|
results: [
|
||||||
|
{
|
||||||
|
id: 11,
|
||||||
|
name: 'group_11',
|
||||||
|
},
|
||||||
|
],
|
||||||
|
},
|
||||||
},
|
},
|
||||||
};
|
};
|
||||||
|
|
||||||
describe('<InventoryHostItem />', () => {
|
describe('<InventoryHostItem />', () => {
|
||||||
let wrapper;
|
const history = createMemoryHistory({
|
||||||
|
initialEntries: ['/inventories/inventory/1/hosts'],
|
||||||
|
});
|
||||||
|
|
||||||
beforeEach(() => {
|
const getChips = (currentScreen) => {
|
||||||
wrapper = mountWithContexts(
|
const list = currentScreen.getByRole('list', {
|
||||||
|
name: 'Related Groups',
|
||||||
|
});
|
||||||
|
const { getAllByRole } = within(list);
|
||||||
|
const items = getAllByRole('listitem');
|
||||||
|
return items.map((item) => item.textContent);
|
||||||
|
};
|
||||||
|
|
||||||
|
const Component = (props) => (
|
||||||
|
<Router history={history}>
|
||||||
<table>
|
<table>
|
||||||
<tbody>
|
<tbody>
|
||||||
<InventoryHostItem
|
<InventoryHostItem
|
||||||
isSelected={false}
|
|
||||||
detailUrl="/host/1"
|
detailUrl="/host/1"
|
||||||
onSelect={() => {}}
|
editUrl={`/inventories/inventory/1/hosts/1/edit`}
|
||||||
host={mockHost}
|
host={mockHost}
|
||||||
|
isSelected={false}
|
||||||
|
onSelect={() => {}}
|
||||||
|
{...props}
|
||||||
/>
|
/>
|
||||||
</tbody>
|
</tbody>
|
||||||
</table>
|
</table>
|
||||||
|
</Router>
|
||||||
);
|
);
|
||||||
|
|
||||||
|
beforeEach(() => {
|
||||||
|
i18n.loadLocaleData({ en: { plurals: en } });
|
||||||
|
i18n.load({ en: english });
|
||||||
|
i18n.activate('en');
|
||||||
});
|
});
|
||||||
|
|
||||||
test('should display expected details', () => {
|
test('should display expected details', () => {
|
||||||
expect(wrapper.find('InventoryHostItem').length).toBe(1);
|
render(<Component />);
|
||||||
expect(wrapper.find('Td[dataLabel="Name"]').find('Link').prop('to')).toBe(
|
|
||||||
|
expect(screen.getByRole('cell', { name: 'Bar' })).toBeInTheDocument();
|
||||||
|
expect(
|
||||||
|
screen.getByRole('checkbox', { name: 'Toggle host' })
|
||||||
|
).toBeInTheDocument();
|
||||||
|
expect(screen.getByRole('link', { name: 'Host 1' })).toHaveAttribute(
|
||||||
|
'href',
|
||||||
'/host/1'
|
'/host/1'
|
||||||
);
|
);
|
||||||
expect(wrapper.find('Td[dataLabel="Description"]').text()).toBe('Bar');
|
expect(screen.getByRole('link', { name: 'Edit host' })).toHaveAttribute(
|
||||||
});
|
'href',
|
||||||
|
'/inventories/inventory/1/hosts/1/edit'
|
||||||
|
);
|
||||||
|
|
||||||
test('edit button shown to users with edit capabilities', () => {
|
const relatedGroupChips = getChips(screen);
|
||||||
expect(wrapper.find('PencilAltIcon').exists()).toBeTruthy();
|
expect(relatedGroupChips).toEqual(['group_11']);
|
||||||
});
|
});
|
||||||
|
|
||||||
test('edit button hidden from users without edit capabilities', () => {
|
test('edit button hidden from users without edit capabilities', () => {
|
||||||
const copyMockHost = { ...mockHost };
|
const copyMockHost = { ...mockHost };
|
||||||
copyMockHost.summary_fields.user_capabilities.edit = false;
|
copyMockHost.summary_fields.user_capabilities.edit = false;
|
||||||
wrapper = mountWithContexts(
|
|
||||||
<table>
|
render(<Component host={copyMockHost} />);
|
||||||
<tbody>
|
expect(screen.queryByText('Edit host')).toBeNull();
|
||||||
<InventoryHostItem
|
|
||||||
isSelected={false}
|
|
||||||
detailUrl="/host/1"
|
|
||||||
onSelect={() => {}}
|
|
||||||
host={copyMockHost}
|
|
||||||
/>
|
|
||||||
</tbody>
|
|
||||||
</table>
|
|
||||||
);
|
|
||||||
expect(wrapper.find('PencilAltIcon').exists()).toBeFalsy();
|
|
||||||
});
|
});
|
||||||
|
|
||||||
test('should display host toggle', () => {
|
test('should show and hide related groups on overflow button click', async () => {
|
||||||
expect(wrapper.find('HostToggle').length).toBe(1);
|
const copyMockHost = { ...mockHost };
|
||||||
|
const mockGroups = [
|
||||||
|
{
|
||||||
|
id: 1,
|
||||||
|
name: 'group_1',
|
||||||
|
},
|
||||||
|
{
|
||||||
|
id: 2,
|
||||||
|
name: 'group_2',
|
||||||
|
},
|
||||||
|
{
|
||||||
|
id: 3,
|
||||||
|
name: 'group_3',
|
||||||
|
},
|
||||||
|
{
|
||||||
|
id: 4,
|
||||||
|
name: 'group_4',
|
||||||
|
},
|
||||||
|
{
|
||||||
|
id: 5,
|
||||||
|
name: 'group_5',
|
||||||
|
},
|
||||||
|
{
|
||||||
|
id: 6,
|
||||||
|
name: 'group_6',
|
||||||
|
},
|
||||||
|
];
|
||||||
|
copyMockHost.summary_fields.groups = {
|
||||||
|
count: 6,
|
||||||
|
results: mockGroups.slice(0, 5),
|
||||||
|
};
|
||||||
|
HostsAPI.readGroups.mockReturnValue({
|
||||||
|
data: {
|
||||||
|
results: mockGroups,
|
||||||
|
},
|
||||||
|
});
|
||||||
|
|
||||||
|
render(<Component host={copyMockHost} />);
|
||||||
|
|
||||||
|
const initialRelatedGroupChips = getChips(screen);
|
||||||
|
expect(initialRelatedGroupChips).toEqual([
|
||||||
|
'group_1',
|
||||||
|
'group_2',
|
||||||
|
'group_3',
|
||||||
|
'group_4',
|
||||||
|
'2 more',
|
||||||
|
]);
|
||||||
|
|
||||||
|
const overflowGroupsButton = screen.queryByText('2 more');
|
||||||
|
fireEvent.click(overflowGroupsButton);
|
||||||
|
|
||||||
|
await waitFor(() => expect(HostsAPI.readGroups).toHaveBeenCalledWith(1));
|
||||||
|
|
||||||
|
const expandedRelatedGroupChips = getChips(screen);
|
||||||
|
expect(expandedRelatedGroupChips).toEqual([
|
||||||
|
'group_1',
|
||||||
|
'group_2',
|
||||||
|
'group_3',
|
||||||
|
'group_4',
|
||||||
|
'group_5',
|
||||||
|
'group_6',
|
||||||
|
'Show less',
|
||||||
|
]);
|
||||||
|
|
||||||
|
const collapseGroupsButton = await screen.findByText('Show less');
|
||||||
|
fireEvent.click(collapseGroupsButton);
|
||||||
|
|
||||||
|
const collapsedRelatedGroupChips = getChips(screen);
|
||||||
|
expect(collapsedRelatedGroupChips).toEqual(initialRelatedGroupChips);
|
||||||
|
});
|
||||||
|
|
||||||
|
test('should show error modal when related groups api request fails', async () => {
|
||||||
|
const copyMockHost = { ...mockHost };
|
||||||
|
const mockGroups = [
|
||||||
|
{
|
||||||
|
id: 1,
|
||||||
|
name: 'group_1',
|
||||||
|
},
|
||||||
|
{
|
||||||
|
id: 2,
|
||||||
|
name: 'group_2',
|
||||||
|
},
|
||||||
|
{
|
||||||
|
id: 3,
|
||||||
|
name: 'group_3',
|
||||||
|
},
|
||||||
|
{
|
||||||
|
id: 4,
|
||||||
|
name: 'group_4',
|
||||||
|
},
|
||||||
|
{
|
||||||
|
id: 5,
|
||||||
|
name: 'group_5',
|
||||||
|
},
|
||||||
|
{
|
||||||
|
id: 6,
|
||||||
|
name: 'group_6',
|
||||||
|
},
|
||||||
|
];
|
||||||
|
copyMockHost.summary_fields.groups = {
|
||||||
|
count: 6,
|
||||||
|
results: mockGroups.slice(0, 5),
|
||||||
|
};
|
||||||
|
HostsAPI.readGroups.mockRejectedValueOnce(new Error());
|
||||||
|
|
||||||
|
render(<Component host={copyMockHost} />);
|
||||||
|
await waitFor(() => {
|
||||||
|
const overflowGroupsButton = screen.queryByText('2 more');
|
||||||
|
fireEvent.click(overflowGroupsButton);
|
||||||
|
});
|
||||||
|
expect(screen.getByRole('dialog', { name: 'Alert modal Error!' }));
|
||||||
});
|
});
|
||||||
});
|
});
|
||||||
|
|||||||
@@ -137,6 +137,7 @@ function InventoryHostList() {
|
|||||||
<HeaderRow qsConfig={QS_CONFIG}>
|
<HeaderRow qsConfig={QS_CONFIG}>
|
||||||
<HeaderCell sortKey="name">{t`Name`}</HeaderCell>
|
<HeaderCell sortKey="name">{t`Name`}</HeaderCell>
|
||||||
<HeaderCell sortKey="description">{t`Description`}</HeaderCell>
|
<HeaderCell sortKey="description">{t`Description`}</HeaderCell>
|
||||||
|
<HeaderCell>{t`Related Groups`}</HeaderCell>
|
||||||
<HeaderCell>{t`Actions`}</HeaderCell>
|
<HeaderCell>{t`Actions`}</HeaderCell>
|
||||||
</HeaderRow>
|
</HeaderRow>
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -41,7 +41,7 @@ function JobEvent({
|
|||||||
if (lineNumber < 0) {
|
if (lineNumber < 0) {
|
||||||
return null;
|
return null;
|
||||||
}
|
}
|
||||||
const canToggle = index === toggleLineIndex;
|
const canToggle = index === toggleLineIndex && !event.isTracebackOnly;
|
||||||
return (
|
return (
|
||||||
<JobEventLine
|
<JobEventLine
|
||||||
onClick={isClickable ? onJobEventClick : undefined}
|
onClick={isClickable ? onJobEventClick : undefined}
|
||||||
@@ -55,7 +55,7 @@ function JobEvent({
|
|||||||
onToggle={onToggleCollapsed}
|
onToggle={onToggleCollapsed}
|
||||||
/>
|
/>
|
||||||
<JobEventLineNumber>
|
<JobEventLineNumber>
|
||||||
{lineNumber}
|
{!event.isTracebackOnly ? lineNumber : ''}
|
||||||
<JobEventEllipsis isCollapsed={isCollapsed && canToggle} />
|
<JobEventEllipsis isCollapsed={isCollapsed && canToggle} />
|
||||||
</JobEventLineNumber>
|
</JobEventLineNumber>
|
||||||
<JobEventLineText
|
<JobEventLineText
|
||||||
|
|||||||
@@ -29,8 +29,11 @@ export function prependTraceback(job, events) {
|
|||||||
start_line: 0,
|
start_line: 0,
|
||||||
};
|
};
|
||||||
const firstIndex = events.findIndex((jobEvent) => jobEvent.counter === 1);
|
const firstIndex = events.findIndex((jobEvent) => jobEvent.counter === 1);
|
||||||
if (firstIndex && events[firstIndex]?.stdout) {
|
if (firstIndex > -1) {
|
||||||
const stdoutLines = events[firstIndex].stdout.split('\r\n');
|
if (!events[firstIndex].stdout) {
|
||||||
|
events[firstIndex].isTracebackOnly = true;
|
||||||
|
}
|
||||||
|
const stdoutLines = events[firstIndex].stdout?.split('\r\n') || [];
|
||||||
stdoutLines[0] = tracebackEvent.stdout;
|
stdoutLines[0] = tracebackEvent.stdout;
|
||||||
events[firstIndex].stdout = stdoutLines.join('\r\n');
|
events[firstIndex].stdout = stdoutLines.join('\r\n');
|
||||||
} else {
|
} else {
|
||||||
|
|||||||
@@ -24,6 +24,7 @@ function WorkflowJobTemplateAdd() {
|
|||||||
limit,
|
limit,
|
||||||
job_tags,
|
job_tags,
|
||||||
skip_tags,
|
skip_tags,
|
||||||
|
scm_branch,
|
||||||
...templatePayload
|
...templatePayload
|
||||||
} = values;
|
} = values;
|
||||||
templatePayload.inventory = inventory?.id;
|
templatePayload.inventory = inventory?.id;
|
||||||
@@ -32,6 +33,7 @@ function WorkflowJobTemplateAdd() {
|
|||||||
templatePayload.limit = limit === '' ? null : limit;
|
templatePayload.limit = limit === '' ? null : limit;
|
||||||
templatePayload.job_tags = job_tags === '' ? null : job_tags;
|
templatePayload.job_tags = job_tags === '' ? null : job_tags;
|
||||||
templatePayload.skip_tags = skip_tags === '' ? null : skip_tags;
|
templatePayload.skip_tags = skip_tags === '' ? null : skip_tags;
|
||||||
|
templatePayload.scm_branch = scm_branch === '' ? null : scm_branch;
|
||||||
const organizationId =
|
const organizationId =
|
||||||
organization?.id || inventory?.summary_fields?.organization.id;
|
organization?.id || inventory?.summary_fields?.organization.id;
|
||||||
try {
|
try {
|
||||||
|
|||||||
@@ -119,7 +119,7 @@ describe('<WorkflowJobTemplateAdd/>', () => {
|
|||||||
job_tags: null,
|
job_tags: null,
|
||||||
limit: null,
|
limit: null,
|
||||||
organization: undefined,
|
organization: undefined,
|
||||||
scm_branch: '',
|
scm_branch: null,
|
||||||
skip_tags: null,
|
skip_tags: null,
|
||||||
webhook_credential: undefined,
|
webhook_credential: undefined,
|
||||||
webhook_service: '',
|
webhook_service: '',
|
||||||
|
|||||||
@@ -30,6 +30,7 @@ function WorkflowJobTemplateEdit({ template }) {
|
|||||||
limit,
|
limit,
|
||||||
job_tags,
|
job_tags,
|
||||||
skip_tags,
|
skip_tags,
|
||||||
|
scm_branch,
|
||||||
...templatePayload
|
...templatePayload
|
||||||
} = values;
|
} = values;
|
||||||
templatePayload.inventory = inventory?.id || null;
|
templatePayload.inventory = inventory?.id || null;
|
||||||
@@ -38,6 +39,7 @@ function WorkflowJobTemplateEdit({ template }) {
|
|||||||
templatePayload.limit = limit === '' ? null : limit;
|
templatePayload.limit = limit === '' ? null : limit;
|
||||||
templatePayload.job_tags = job_tags === '' ? null : job_tags;
|
templatePayload.job_tags = job_tags === '' ? null : job_tags;
|
||||||
templatePayload.skip_tags = skip_tags === '' ? null : skip_tags;
|
templatePayload.skip_tags = skip_tags === '' ? null : skip_tags;
|
||||||
|
templatePayload.scm_branch = scm_branch === '' ? null : scm_branch;
|
||||||
|
|
||||||
const formOrgId =
|
const formOrgId =
|
||||||
organization?.id || inventory?.summary_fields?.organization.id || null;
|
organization?.id || inventory?.summary_fields?.organization.id || null;
|
||||||
|
|||||||
@@ -7,7 +7,6 @@ __metaclass__ = type
|
|||||||
|
|
||||||
DOCUMENTATION = '''
|
DOCUMENTATION = '''
|
||||||
name: controller
|
name: controller
|
||||||
plugin_type: inventory
|
|
||||||
author:
|
author:
|
||||||
- Matthew Jones (@matburt)
|
- Matthew Jones (@matburt)
|
||||||
- Yunfan Zhang (@YunfanZhang42)
|
- Yunfan Zhang (@YunfanZhang42)
|
||||||
|
|||||||
@@ -5,7 +5,7 @@ from __future__ import absolute_import, division, print_function
|
|||||||
__metaclass__ = type
|
__metaclass__ = type
|
||||||
|
|
||||||
DOCUMENTATION = """
|
DOCUMENTATION = """
|
||||||
lookup: controller_api
|
name: controller_api
|
||||||
author: John Westcott IV (@john-westcott-iv)
|
author: John Westcott IV (@john-westcott-iv)
|
||||||
short_description: Search the API for objects
|
short_description: Search the API for objects
|
||||||
requirements:
|
requirements:
|
||||||
@@ -74,7 +74,7 @@ EXAMPLES = """
|
|||||||
|
|
||||||
- name: Load the UI settings specifying the connection info
|
- name: Load the UI settings specifying the connection info
|
||||||
set_fact:
|
set_fact:
|
||||||
controller_settings: "{{ lookup('awx.awx.controller_api', 'settings/ui' host='controller.example.com',
|
controller_settings: "{{ lookup('awx.awx.controller_api', 'settings/ui', host='controller.example.com',
|
||||||
username='admin', password=my_pass_var, verify_ssl=False) }}"
|
username='admin', password=my_pass_var, verify_ssl=False) }}"
|
||||||
|
|
||||||
- name: Report the usernames of all users with admin privs
|
- name: Report the usernames of all users with admin privs
|
||||||
|
|||||||
@@ -5,7 +5,7 @@ from __future__ import absolute_import, division, print_function
|
|||||||
__metaclass__ = type
|
__metaclass__ = type
|
||||||
|
|
||||||
DOCUMENTATION = """
|
DOCUMENTATION = """
|
||||||
lookup: schedule_rrule
|
name: schedule_rrule
|
||||||
author: John Westcott IV (@john-westcott-iv)
|
author: John Westcott IV (@john-westcott-iv)
|
||||||
short_description: Generate an rrule string which can be used for Schedules
|
short_description: Generate an rrule string which can be used for Schedules
|
||||||
requirements:
|
requirements:
|
||||||
@@ -101,7 +101,13 @@ else:
|
|||||||
|
|
||||||
|
|
||||||
class LookupModule(LookupBase):
|
class LookupModule(LookupBase):
|
||||||
frequencies = {
|
# plugin constructor
|
||||||
|
def __init__(self, *args, **kwargs):
|
||||||
|
if LIBRARY_IMPORT_ERROR:
|
||||||
|
raise_from(AnsibleError('{0}'.format(LIBRARY_IMPORT_ERROR)), LIBRARY_IMPORT_ERROR)
|
||||||
|
super().__init__(*args, **kwargs)
|
||||||
|
|
||||||
|
self.frequencies = {
|
||||||
'none': rrule.DAILY,
|
'none': rrule.DAILY,
|
||||||
'minute': rrule.MINUTELY,
|
'minute': rrule.MINUTELY,
|
||||||
'hour': rrule.HOURLY,
|
'hour': rrule.HOURLY,
|
||||||
@@ -110,7 +116,7 @@ class LookupModule(LookupBase):
|
|||||||
'month': rrule.MONTHLY,
|
'month': rrule.MONTHLY,
|
||||||
}
|
}
|
||||||
|
|
||||||
weekdays = {
|
self.weekdays = {
|
||||||
'monday': rrule.MO,
|
'monday': rrule.MO,
|
||||||
'tuesday': rrule.TU,
|
'tuesday': rrule.TU,
|
||||||
'wednesday': rrule.WE,
|
'wednesday': rrule.WE,
|
||||||
@@ -120,7 +126,7 @@ class LookupModule(LookupBase):
|
|||||||
'sunday': rrule.SU,
|
'sunday': rrule.SU,
|
||||||
}
|
}
|
||||||
|
|
||||||
set_positions = {
|
self.set_positions = {
|
||||||
'first': 1,
|
'first': 1,
|
||||||
'second': 2,
|
'second': 2,
|
||||||
'third': 3,
|
'third': 3,
|
||||||
@@ -128,12 +134,6 @@ class LookupModule(LookupBase):
|
|||||||
'last': -1,
|
'last': -1,
|
||||||
}
|
}
|
||||||
|
|
||||||
# plugin constructor
|
|
||||||
def __init__(self, *args, **kwargs):
|
|
||||||
if LIBRARY_IMPORT_ERROR:
|
|
||||||
raise_from(AnsibleError('{0}'.format(LIBRARY_IMPORT_ERROR)), LIBRARY_IMPORT_ERROR)
|
|
||||||
super().__init__(*args, **kwargs)
|
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def parse_date_time(date_string):
|
def parse_date_time(date_string):
|
||||||
try:
|
try:
|
||||||
@@ -149,14 +149,13 @@ class LookupModule(LookupBase):
|
|||||||
|
|
||||||
return self.get_rrule(frequency, kwargs)
|
return self.get_rrule(frequency, kwargs)
|
||||||
|
|
||||||
@staticmethod
|
def get_rrule(self, frequency, kwargs):
|
||||||
def get_rrule(frequency, kwargs):
|
|
||||||
|
|
||||||
if frequency not in LookupModule.frequencies:
|
if frequency not in self.frequencies:
|
||||||
raise AnsibleError('Frequency of {0} is invalid'.format(frequency))
|
raise AnsibleError('Frequency of {0} is invalid'.format(frequency))
|
||||||
|
|
||||||
rrule_kwargs = {
|
rrule_kwargs = {
|
||||||
'freq': LookupModule.frequencies[frequency],
|
'freq': self.frequencies[frequency],
|
||||||
'interval': kwargs.get('every', 1),
|
'interval': kwargs.get('every', 1),
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -187,9 +186,9 @@ class LookupModule(LookupBase):
|
|||||||
days = []
|
days = []
|
||||||
for day in kwargs['on_days'].split(','):
|
for day in kwargs['on_days'].split(','):
|
||||||
day = day.strip()
|
day = day.strip()
|
||||||
if day not in LookupModule.weekdays:
|
if day not in self.weekdays:
|
||||||
raise AnsibleError('Parameter on_days must only contain values {0}'.format(', '.join(LookupModule.weekdays.keys())))
|
raise AnsibleError('Parameter on_days must only contain values {0}'.format(', '.join(self.weekdays.keys())))
|
||||||
days.append(LookupModule.weekdays[day])
|
days.append(self.weekdays[day])
|
||||||
|
|
||||||
rrule_kwargs['byweekday'] = days
|
rrule_kwargs['byweekday'] = days
|
||||||
|
|
||||||
@@ -214,13 +213,13 @@ class LookupModule(LookupBase):
|
|||||||
except Exception as e:
|
except Exception as e:
|
||||||
raise_from(AnsibleError('on_the parameter must be two words separated by a space'), e)
|
raise_from(AnsibleError('on_the parameter must be two words separated by a space'), e)
|
||||||
|
|
||||||
if weekday not in LookupModule.weekdays:
|
if weekday not in self.weekdays:
|
||||||
raise AnsibleError('Weekday portion of on_the parameter is not valid')
|
raise AnsibleError('Weekday portion of on_the parameter is not valid')
|
||||||
if occurance not in LookupModule.set_positions:
|
if occurance not in self.set_positions:
|
||||||
raise AnsibleError('The first string of the on_the parameter is not valid')
|
raise AnsibleError('The first string of the on_the parameter is not valid')
|
||||||
|
|
||||||
rrule_kwargs['byweekday'] = LookupModule.weekdays[weekday]
|
rrule_kwargs['byweekday'] = self.weekdays[weekday]
|
||||||
rrule_kwargs['bysetpos'] = LookupModule.set_positions[occurance]
|
rrule_kwargs['bysetpos'] = self.set_positions[occurance]
|
||||||
|
|
||||||
my_rule = rrule.rrule(**rrule_kwargs)
|
my_rule = rrule.rrule(**rrule_kwargs)
|
||||||
|
|
||||||
|
|||||||
@@ -5,7 +5,7 @@ from __future__ import absolute_import, division, print_function
|
|||||||
__metaclass__ = type
|
__metaclass__ = type
|
||||||
|
|
||||||
DOCUMENTATION = """
|
DOCUMENTATION = """
|
||||||
lookup: schedule_rruleset
|
name: schedule_rruleset
|
||||||
author: John Westcott IV (@john-westcott-iv)
|
author: John Westcott IV (@john-westcott-iv)
|
||||||
short_description: Generate an rruleset string
|
short_description: Generate an rruleset string
|
||||||
requirements:
|
requirements:
|
||||||
@@ -31,7 +31,8 @@ DOCUMENTATION = """
|
|||||||
rules:
|
rules:
|
||||||
description:
|
description:
|
||||||
- Array of rules in the rruleset
|
- Array of rules in the rruleset
|
||||||
type: array
|
type: list
|
||||||
|
elements: dict
|
||||||
required: True
|
required: True
|
||||||
suboptions:
|
suboptions:
|
||||||
frequency:
|
frequency:
|
||||||
@@ -136,11 +137,19 @@ try:
|
|||||||
import pytz
|
import pytz
|
||||||
from dateutil import rrule
|
from dateutil import rrule
|
||||||
except ImportError as imp_exc:
|
except ImportError as imp_exc:
|
||||||
raise_from(AnsibleError('{0}'.format(imp_exc)), imp_exc)
|
LIBRARY_IMPORT_ERROR = imp_exc
|
||||||
|
else:
|
||||||
|
LIBRARY_IMPORT_ERROR = None
|
||||||
|
|
||||||
|
|
||||||
class LookupModule(LookupBase):
|
class LookupModule(LookupBase):
|
||||||
frequencies = {
|
# plugin constructor
|
||||||
|
def __init__(self, *args, **kwargs):
|
||||||
|
if LIBRARY_IMPORT_ERROR:
|
||||||
|
raise_from(AnsibleError('{0}'.format(LIBRARY_IMPORT_ERROR)), LIBRARY_IMPORT_ERROR)
|
||||||
|
super().__init__(*args, **kwargs)
|
||||||
|
|
||||||
|
self.frequencies = {
|
||||||
'none': rrule.DAILY,
|
'none': rrule.DAILY,
|
||||||
'minute': rrule.MINUTELY,
|
'minute': rrule.MINUTELY,
|
||||||
'hour': rrule.HOURLY,
|
'hour': rrule.HOURLY,
|
||||||
@@ -149,7 +158,7 @@ class LookupModule(LookupBase):
|
|||||||
'month': rrule.MONTHLY,
|
'month': rrule.MONTHLY,
|
||||||
}
|
}
|
||||||
|
|
||||||
weekdays = {
|
self.weekdays = {
|
||||||
'monday': rrule.MO,
|
'monday': rrule.MO,
|
||||||
'tuesday': rrule.TU,
|
'tuesday': rrule.TU,
|
||||||
'wednesday': rrule.WE,
|
'wednesday': rrule.WE,
|
||||||
@@ -159,7 +168,7 @@ class LookupModule(LookupBase):
|
|||||||
'sunday': rrule.SU,
|
'sunday': rrule.SU,
|
||||||
}
|
}
|
||||||
|
|
||||||
set_positions = {
|
self.set_positions = {
|
||||||
'first': 1,
|
'first': 1,
|
||||||
'second': 2,
|
'second': 2,
|
||||||
'third': 3,
|
'third': 3,
|
||||||
@@ -167,10 +176,6 @@ class LookupModule(LookupBase):
|
|||||||
'last': -1,
|
'last': -1,
|
||||||
}
|
}
|
||||||
|
|
||||||
# plugin constructor
|
|
||||||
def __init__(self, *args, **kwargs):
|
|
||||||
super().__init__(*args, **kwargs)
|
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def parse_date_time(date_string):
|
def parse_date_time(date_string):
|
||||||
try:
|
try:
|
||||||
@@ -188,14 +193,14 @@ class LookupModule(LookupBase):
|
|||||||
# something: [1,2,3] - A list of ints
|
# something: [1,2,3] - A list of ints
|
||||||
return_values = []
|
return_values = []
|
||||||
# If they give us a single int, lets make it a list of ints
|
# If they give us a single int, lets make it a list of ints
|
||||||
if type(rule[field_name]) == int:
|
if isinstance(rule[field_name], int):
|
||||||
rule[field_name] = [rule[field_name]]
|
rule[field_name] = [rule[field_name]]
|
||||||
# If its not a list, we need to split it into a list
|
# If its not a list, we need to split it into a list
|
||||||
if type(rule[field_name]) != list:
|
if isinstance(rule[field_name], list):
|
||||||
rule[field_name] = rule[field_name].split(',')
|
rule[field_name] = rule[field_name].split(',')
|
||||||
for value in rule[field_name]:
|
for value in rule[field_name]:
|
||||||
# If they have a list of strs we want to strip the str incase its space delineated
|
# If they have a list of strs we want to strip the str incase its space delineated
|
||||||
if type(value) == str:
|
if isinstance(value, str):
|
||||||
value = value.strip()
|
value = value.strip()
|
||||||
# If value happens to be an int (from a list of ints) we need to coerce it into a str for the re.match
|
# If value happens to be an int (from a list of ints) we need to coerce it into a str for the re.match
|
||||||
if not re.match(r"^\d+$", str(value)) or int(value) < min_value or int(value) > max_value:
|
if not re.match(r"^\d+$", str(value)) or int(value) < min_value or int(value) > max_value:
|
||||||
@@ -205,7 +210,7 @@ class LookupModule(LookupBase):
|
|||||||
|
|
||||||
def process_list(self, field_name, rule, valid_list, rule_number):
|
def process_list(self, field_name, rule, valid_list, rule_number):
|
||||||
return_values = []
|
return_values = []
|
||||||
if type(rule[field_name]) != list:
|
if isinstance(rule[field_name], list):
|
||||||
rule[field_name] = rule[field_name].split(',')
|
rule[field_name] = rule[field_name].split(',')
|
||||||
for value in rule[field_name]:
|
for value in rule[field_name]:
|
||||||
value = value.strip()
|
value = value.strip()
|
||||||
@@ -260,11 +265,11 @@ class LookupModule(LookupBase):
|
|||||||
frequency = rule.get('frequency', None)
|
frequency = rule.get('frequency', None)
|
||||||
if not frequency:
|
if not frequency:
|
||||||
raise AnsibleError("Rule {0} is missing a frequency".format(rule_number))
|
raise AnsibleError("Rule {0} is missing a frequency".format(rule_number))
|
||||||
if frequency not in LookupModule.frequencies:
|
if frequency not in self.frequencies:
|
||||||
raise AnsibleError('Frequency of rule {0} is invalid {1}'.format(rule_number, frequency))
|
raise AnsibleError('Frequency of rule {0} is invalid {1}'.format(rule_number, frequency))
|
||||||
|
|
||||||
rrule_kwargs = {
|
rrule_kwargs = {
|
||||||
'freq': LookupModule.frequencies[frequency],
|
'freq': self.frequencies[frequency],
|
||||||
'interval': rule.get('interval', 1),
|
'interval': rule.get('interval', 1),
|
||||||
'dtstart': start_date,
|
'dtstart': start_date,
|
||||||
}
|
}
|
||||||
@@ -287,7 +292,7 @@ class LookupModule(LookupBase):
|
|||||||
)
|
)
|
||||||
|
|
||||||
if 'bysetpos' in rule:
|
if 'bysetpos' in rule:
|
||||||
rrule_kwargs['bysetpos'] = self.process_list('bysetpos', rule, LookupModule.set_positions, rule_number)
|
rrule_kwargs['bysetpos'] = self.process_list('bysetpos', rule, self.set_positions, rule_number)
|
||||||
|
|
||||||
if 'bymonth' in rule:
|
if 'bymonth' in rule:
|
||||||
rrule_kwargs['bymonth'] = self.process_integer('bymonth', rule, 1, 12, rule_number)
|
rrule_kwargs['bymonth'] = self.process_integer('bymonth', rule, 1, 12, rule_number)
|
||||||
@@ -302,7 +307,7 @@ class LookupModule(LookupBase):
|
|||||||
rrule_kwargs['byweekno'] = self.process_integer('byweekno', rule, 1, 52, rule_number)
|
rrule_kwargs['byweekno'] = self.process_integer('byweekno', rule, 1, 52, rule_number)
|
||||||
|
|
||||||
if 'byweekday' in rule:
|
if 'byweekday' in rule:
|
||||||
rrule_kwargs['byweekday'] = self.process_list('byweekday', rule, LookupModule.weekdays, rule_number)
|
rrule_kwargs['byweekday'] = self.process_list('byweekday', rule, self.weekdays, rule_number)
|
||||||
|
|
||||||
if 'byhour' in rule:
|
if 'byhour' in rule:
|
||||||
rrule_kwargs['byhour'] = self.process_integer('byhour', rule, 0, 23, rule_number)
|
rrule_kwargs['byhour'] = self.process_integer('byhour', rule, 0, 23, rule_number)
|
||||||
|
|||||||
@@ -4,6 +4,7 @@ __metaclass__ = type
|
|||||||
|
|
||||||
from ansible.module_utils.basic import AnsibleModule, env_fallback
|
from ansible.module_utils.basic import AnsibleModule, env_fallback
|
||||||
from ansible.module_utils.urls import Request, SSLValidationError, ConnectionError
|
from ansible.module_utils.urls import Request, SSLValidationError, ConnectionError
|
||||||
|
from ansible.module_utils.parsing.convert_bool import boolean as strtobool
|
||||||
from ansible.module_utils.six import PY2
|
from ansible.module_utils.six import PY2
|
||||||
from ansible.module_utils.six import raise_from, string_types
|
from ansible.module_utils.six import raise_from, string_types
|
||||||
from ansible.module_utils.six.moves import StringIO
|
from ansible.module_utils.six.moves import StringIO
|
||||||
@@ -11,14 +12,21 @@ from ansible.module_utils.six.moves.urllib.error import HTTPError
|
|||||||
from ansible.module_utils.six.moves.http_cookiejar import CookieJar
|
from ansible.module_utils.six.moves.http_cookiejar import CookieJar
|
||||||
from ansible.module_utils.six.moves.urllib.parse import urlparse, urlencode
|
from ansible.module_utils.six.moves.urllib.parse import urlparse, urlencode
|
||||||
from ansible.module_utils.six.moves.configparser import ConfigParser, NoOptionError
|
from ansible.module_utils.six.moves.configparser import ConfigParser, NoOptionError
|
||||||
from distutils.version import LooseVersion as Version
|
|
||||||
from socket import getaddrinfo, IPPROTO_TCP
|
from socket import getaddrinfo, IPPROTO_TCP
|
||||||
import time
|
import time
|
||||||
import re
|
import re
|
||||||
from json import loads, dumps
|
from json import loads, dumps
|
||||||
from os.path import isfile, expanduser, split, join, exists, isdir
|
from os.path import isfile, expanduser, split, join, exists, isdir
|
||||||
from os import access, R_OK, getcwd
|
from os import access, R_OK, getcwd
|
||||||
from distutils.util import strtobool
|
|
||||||
|
|
||||||
|
try:
|
||||||
|
from ansible.module_utils.compat.version import LooseVersion as Version
|
||||||
|
except ImportError:
|
||||||
|
try:
|
||||||
|
from distutils.version import LooseVersion as Version
|
||||||
|
except ImportError:
|
||||||
|
raise AssertionError('To use this plugin or module with ansible-core 2.11, you need to use Python < 3.12 with distutils.version present')
|
||||||
|
|
||||||
try:
|
try:
|
||||||
import yaml
|
import yaml
|
||||||
|
|||||||
@@ -55,7 +55,6 @@ options:
|
|||||||
description:
|
description:
|
||||||
- The arguments to pass to the module.
|
- The arguments to pass to the module.
|
||||||
type: str
|
type: str
|
||||||
default: ""
|
|
||||||
forks:
|
forks:
|
||||||
description:
|
description:
|
||||||
- The number of forks to use for this ad hoc execution.
|
- The number of forks to use for this ad hoc execution.
|
||||||
|
|||||||
@@ -42,6 +42,7 @@ options:
|
|||||||
- Maximum time in seconds to wait for a job to finish.
|
- Maximum time in seconds to wait for a job to finish.
|
||||||
- Not specifying means the task will wait until the controller cancels the command.
|
- Not specifying means the task will wait until the controller cancels the command.
|
||||||
type: int
|
type: int
|
||||||
|
default: 0
|
||||||
extends_documentation_fragment: awx.awx.auth
|
extends_documentation_fragment: awx.awx.auth
|
||||||
'''
|
'''
|
||||||
|
|
||||||
|
|||||||
@@ -52,7 +52,7 @@ options:
|
|||||||
- The credential type being created.
|
- The credential type being created.
|
||||||
- Can be a built-in credential type such as "Machine", or a custom credential type such as "My Credential Type"
|
- Can be a built-in credential type such as "Machine", or a custom credential type such as "My Credential Type"
|
||||||
- Choices include Amazon Web Services, Ansible Galaxy/Automation Hub API Token, Centrify Vault Credential Provider Lookup,
|
- Choices include Amazon Web Services, Ansible Galaxy/Automation Hub API Token, Centrify Vault Credential Provider Lookup,
|
||||||
Container Registry, CyberArk AIM Central Credential Provider Lookup, CyberArk Conjur Secrets Manager Lookup, Google Compute Engine,
|
Container Registry, CyberArk Central Credential Provider Lookup, CyberArk Conjur Secret Lookup, Google Compute Engine,
|
||||||
GitHub Personal Access Token, GitLab Personal Access Token, GPG Public Key, HashiCorp Vault Secret Lookup, HashiCorp Vault Signed SSH,
|
GitHub Personal Access Token, GitLab Personal Access Token, GPG Public Key, HashiCorp Vault Secret Lookup, HashiCorp Vault Signed SSH,
|
||||||
Insights, Machine, Microsoft Azure Key Vault, Microsoft Azure Resource Manager, Network, OpenShift or Kubernetes API
|
Insights, Machine, Microsoft Azure Key Vault, Microsoft Azure Resource Manager, Network, OpenShift or Kubernetes API
|
||||||
Bearer Token, OpenStack, Red Hat Ansible Automation Platform, Red Hat Satellite 6, Red Hat Virtualization, Source Control,
|
Bearer Token, OpenStack, Red Hat Ansible Automation Platform, Red Hat Satellite 6, Red Hat Virtualization, Source Control,
|
||||||
|
|||||||
@@ -80,9 +80,9 @@ def main():
|
|||||||
name=dict(required=True),
|
name=dict(required=True),
|
||||||
new_name=dict(),
|
new_name=dict(),
|
||||||
image=dict(required=True),
|
image=dict(required=True),
|
||||||
description=dict(default=''),
|
description=dict(),
|
||||||
organization=dict(),
|
organization=dict(),
|
||||||
credential=dict(default=''),
|
credential=dict(),
|
||||||
state=dict(choices=['present', 'absent'], default='present'),
|
state=dict(choices=['present', 'absent'], default='present'),
|
||||||
pull=dict(choices=['always', 'missing', 'never'], default='missing'),
|
pull=dict(choices=['always', 'missing', 'never'], default='missing'),
|
||||||
)
|
)
|
||||||
|
|||||||
@@ -28,52 +28,74 @@ options:
|
|||||||
default: 'False'
|
default: 'False'
|
||||||
organizations:
|
organizations:
|
||||||
description:
|
description:
|
||||||
- organization name to export
|
- organization names to export
|
||||||
type: str
|
type: list
|
||||||
|
elements: str
|
||||||
users:
|
users:
|
||||||
description:
|
description:
|
||||||
- user name to export
|
- user names to export
|
||||||
type: str
|
type: list
|
||||||
|
elements: str
|
||||||
teams:
|
teams:
|
||||||
description:
|
description:
|
||||||
- team name to export
|
- team names to export
|
||||||
type: str
|
type: list
|
||||||
|
elements: str
|
||||||
credential_types:
|
credential_types:
|
||||||
description:
|
description:
|
||||||
- credential type name to export
|
- credential type names to export
|
||||||
type: str
|
type: list
|
||||||
|
elements: str
|
||||||
credentials:
|
credentials:
|
||||||
description:
|
description:
|
||||||
- credential name to export
|
- credential names to export
|
||||||
type: str
|
type: list
|
||||||
|
elements: str
|
||||||
execution_environments:
|
execution_environments:
|
||||||
description:
|
description:
|
||||||
- execution environment name to export
|
- execution environment names to export
|
||||||
type: str
|
type: list
|
||||||
|
elements: str
|
||||||
notification_templates:
|
notification_templates:
|
||||||
description:
|
description:
|
||||||
- notification template name to export
|
- notification template names to export
|
||||||
type: str
|
type: list
|
||||||
|
elements: str
|
||||||
inventory_sources:
|
inventory_sources:
|
||||||
description:
|
description:
|
||||||
- inventory soruce to export
|
- inventory soruces to export
|
||||||
type: str
|
type: list
|
||||||
|
elements: str
|
||||||
inventory:
|
inventory:
|
||||||
description:
|
description:
|
||||||
- inventory name to export
|
- inventory names to export
|
||||||
type: str
|
type: list
|
||||||
|
elements: str
|
||||||
projects:
|
projects:
|
||||||
description:
|
description:
|
||||||
- project name to export
|
- project names to export
|
||||||
type: str
|
type: list
|
||||||
|
elements: str
|
||||||
job_templates:
|
job_templates:
|
||||||
description:
|
description:
|
||||||
- job template name to export
|
- job template names to export
|
||||||
type: str
|
type: list
|
||||||
|
elements: str
|
||||||
workflow_job_templates:
|
workflow_job_templates:
|
||||||
description:
|
description:
|
||||||
- workflow name to export
|
- workflow names to export
|
||||||
type: str
|
type: list
|
||||||
|
elements: str
|
||||||
|
applications:
|
||||||
|
description:
|
||||||
|
- OAuth2 application names to export
|
||||||
|
type: list
|
||||||
|
elements: str
|
||||||
|
schedules:
|
||||||
|
description:
|
||||||
|
- schedule names to export
|
||||||
|
type: list
|
||||||
|
elements: str
|
||||||
requirements:
|
requirements:
|
||||||
- "awxkit >= 9.3.0"
|
- "awxkit >= 9.3.0"
|
||||||
notes:
|
notes:
|
||||||
@@ -94,6 +116,10 @@ EXAMPLES = '''
|
|||||||
export:
|
export:
|
||||||
job_templates: "My Template"
|
job_templates: "My Template"
|
||||||
credential: 'all'
|
credential: 'all'
|
||||||
|
|
||||||
|
- name: Export a list of inventories
|
||||||
|
export:
|
||||||
|
inventory: ['My Inventory 1', 'My Inventory 2']
|
||||||
'''
|
'''
|
||||||
|
|
||||||
import logging
|
import logging
|
||||||
@@ -111,24 +137,12 @@ except ImportError:
|
|||||||
def main():
|
def main():
|
||||||
argument_spec = dict(
|
argument_spec = dict(
|
||||||
all=dict(type='bool', default=False),
|
all=dict(type='bool', default=False),
|
||||||
credential_types=dict(type='str'),
|
|
||||||
credentials=dict(type='str'),
|
|
||||||
execution_environments=dict(type='str'),
|
|
||||||
inventory=dict(type='str'),
|
|
||||||
inventory_sources=dict(type='str'),
|
|
||||||
job_templates=dict(type='str'),
|
|
||||||
notification_templates=dict(type='str'),
|
|
||||||
organizations=dict(type='str'),
|
|
||||||
projects=dict(type='str'),
|
|
||||||
teams=dict(type='str'),
|
|
||||||
users=dict(type='str'),
|
|
||||||
workflow_job_templates=dict(type='str'),
|
|
||||||
)
|
)
|
||||||
|
|
||||||
# We are not going to raise an error here because the __init__ method of ControllerAWXKitModule will do that for us
|
# We are not going to raise an error here because the __init__ method of ControllerAWXKitModule will do that for us
|
||||||
if HAS_EXPORTABLE_RESOURCES:
|
if HAS_EXPORTABLE_RESOURCES:
|
||||||
for resource in EXPORTABLE_RESOURCES:
|
for resource in EXPORTABLE_RESOURCES:
|
||||||
argument_spec[resource] = dict(type='str')
|
argument_spec[resource] = dict(type='list', elements='str')
|
||||||
|
|
||||||
module = ControllerAWXKitModule(argument_spec=argument_spec)
|
module = ControllerAWXKitModule(argument_spec=argument_spec)
|
||||||
|
|
||||||
|
|||||||
@@ -54,6 +54,18 @@ options:
|
|||||||
required: False
|
required: False
|
||||||
type: int
|
type: int
|
||||||
default: '0'
|
default: '0'
|
||||||
|
max_concurrent_jobs:
|
||||||
|
description:
|
||||||
|
- Maximum number of concurrent jobs to run on this group. Zero means no limit.
|
||||||
|
required: False
|
||||||
|
type: int
|
||||||
|
default: '0'
|
||||||
|
max_forks:
|
||||||
|
description:
|
||||||
|
- Max forks to execute on this group. Zero means no limit.
|
||||||
|
required: False
|
||||||
|
type: int
|
||||||
|
default: '0'
|
||||||
policy_instance_list:
|
policy_instance_list:
|
||||||
description:
|
description:
|
||||||
- List of exact-match Instances that will be assigned to this group
|
- List of exact-match Instances that will be assigned to this group
|
||||||
@@ -95,6 +107,8 @@ def main():
|
|||||||
is_container_group=dict(type='bool', default=False),
|
is_container_group=dict(type='bool', default=False),
|
||||||
policy_instance_percentage=dict(type='int', default='0'),
|
policy_instance_percentage=dict(type='int', default='0'),
|
||||||
policy_instance_minimum=dict(type='int', default='0'),
|
policy_instance_minimum=dict(type='int', default='0'),
|
||||||
|
max_concurrent_jobs=dict(type='int', default='0'),
|
||||||
|
max_forks=dict(type='int', default='0'),
|
||||||
policy_instance_list=dict(type='list', elements='str'),
|
policy_instance_list=dict(type='list', elements='str'),
|
||||||
pod_spec_override=dict(),
|
pod_spec_override=dict(),
|
||||||
instances=dict(required=False, type="list", elements='str', default=None),
|
instances=dict(required=False, type="list", elements='str', default=None),
|
||||||
@@ -111,6 +125,8 @@ def main():
|
|||||||
is_container_group = module.params.get('is_container_group')
|
is_container_group = module.params.get('is_container_group')
|
||||||
policy_instance_percentage = module.params.get('policy_instance_percentage')
|
policy_instance_percentage = module.params.get('policy_instance_percentage')
|
||||||
policy_instance_minimum = module.params.get('policy_instance_minimum')
|
policy_instance_minimum = module.params.get('policy_instance_minimum')
|
||||||
|
max_concurrent_jobs = module.params.get('max_concurrent_jobs')
|
||||||
|
max_forks = module.params.get('max_forks')
|
||||||
policy_instance_list = module.params.get('policy_instance_list')
|
policy_instance_list = module.params.get('policy_instance_list')
|
||||||
pod_spec_override = module.params.get('pod_spec_override')
|
pod_spec_override = module.params.get('pod_spec_override')
|
||||||
instances = module.params.get('instances')
|
instances = module.params.get('instances')
|
||||||
@@ -144,6 +160,10 @@ def main():
|
|||||||
new_fields['policy_instance_percentage'] = policy_instance_percentage
|
new_fields['policy_instance_percentage'] = policy_instance_percentage
|
||||||
if policy_instance_minimum is not None:
|
if policy_instance_minimum is not None:
|
||||||
new_fields['policy_instance_minimum'] = policy_instance_minimum
|
new_fields['policy_instance_minimum'] = policy_instance_minimum
|
||||||
|
if max_concurrent_jobs is not None:
|
||||||
|
new_fields['max_concurrent_jobs'] = max_concurrent_jobs
|
||||||
|
if max_forks is not None:
|
||||||
|
new_fields['max_forks'] = max_forks
|
||||||
if policy_instance_list is not None:
|
if policy_instance_list is not None:
|
||||||
new_fields['policy_instance_list'] = policy_instance_list
|
new_fields['policy_instance_list'] = policy_instance_list
|
||||||
if pod_spec_override is not None:
|
if pod_spec_override is not None:
|
||||||
|
|||||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user