Compare commits

..

1151 Commits
3.0.1 ... 5.0.0

Author SHA1 Message Date
Shane McDonald
08aff9bd2c AWX 5.0.0 2019-06-18 10:12:42 -04:00
softwarefactory-project-zuul[bot]
1897b18a6e Merge pull request #4090 from Spredzy/bump_pip
Makefile: align pip and setuptools bump

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-06-17 20:52:44 +00:00
Yanis Guenane
ec92abf014 Makefile: align pip and setuptools bump
Signed-off-by: Yanis Guenane <yanis@guenane.org>
2019-06-17 21:58:42 +02:00
softwarefactory-project-zuul[bot]
5ed7db8cc2 Merge pull request #4085 from mabashian/4082-inv_obj
Ensure inventory_obj is on scope before checking it

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-06-17 18:08:43 +00:00
mabashian
863d962ec2 Ensure inventory_obj is on scope before checking it 2019-06-17 07:40:08 -04:00
softwarefactory-project-zuul[bot]
049b3a2e87 Merge pull request #4080 from AlanCoding/no_more_v1
Remove code used defunct jobs list POST

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-06-15 01:02:24 +00:00
AlanCoding
196b6572b2 remove code used defunct jobs list POST 2019-06-14 13:13:37 -04:00
softwarefactory-project-zuul[bot]
37cb912367 Merge pull request #4075 from ryanpetrello/old-migration-cleanup
delete a few really old migration utility files

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-06-14 14:27:47 +00:00
softwarefactory-project-zuul[bot]
9958f382d7 Merge pull request #4043 from mabashian/4029-inv-refresh
Fixes bug where the delete inventory event would wipe or reset inventory form fields

Reviewed-by: Michael Abashian
             https://github.com/mabashian
2019-06-14 14:04:59 +00:00
softwarefactory-project-zuul[bot]
1d767a15d8 Merge pull request #4044 from mabashian/read-only-codemirrors
Read-only codemirror cleanup

Reviewed-by: Michael Abashian
             https://github.com/mabashian
2019-06-14 13:58:51 +00:00
softwarefactory-project-zuul[bot]
612373c849 Merge pull request #4041 from mabashian/wf-execute-unsaved
Fixes bug where save changes modal was shown to user with execute permissions

Reviewed-by: Michael Abashian
             https://github.com/mabashian
2019-06-14 13:45:33 +00:00
mabashian
3ea6171b54 Fixes bug where the delete inventory event would wipe or reset inventory form fields 2019-06-14 09:38:27 -04:00
mabashian
302d8589c9 Make extra vars read-only for users without edit role on some forms. Fixes console error thrown in launch modal if survey tab missing. 2019-06-14 09:37:37 -04:00
mabashian
01f51219a6 Fix linting errors 2019-06-14 09:22:16 -04:00
mabashian
38ea82bf3d Fixes bug where save changes modal was shown to user with execute permissions 2019-06-14 09:22:16 -04:00
softwarefactory-project-zuul[bot]
d4ad674899 Merge pull request #4064 from mabashian/3829-new-list-links
Fixes links in various lists to maintain query params while navigating to edit form

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-06-13 21:56:34 +00:00
softwarefactory-project-zuul[bot]
9d3aca5e1b Merge pull request #4060 from mabashian/3830-host-filter-search
Always show search bar in smart inventory host filter modal

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-06-13 21:44:02 +00:00
Ryan Petrello
a454102e77 delete a few really old migration utility files 2019-06-13 17:19:43 -04:00
softwarefactory-project-zuul[bot]
e1d60ff4f1 Merge pull request #4052 from gamuniz/enabled_value_fix
Enabled value fix to be case insensitive(GCE API)

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-06-13 20:08:30 +00:00
Gabe Muniz
bd93d97a60 fix so enabled value is case insensitive 2019-06-13 15:36:54 -04:00
softwarefactory-project-zuul[bot]
793e78d9c0 Merge pull request #4046 from AlanCoding/setuptools
Upgrade pip and setuptools

Reviewed-by: Alan Rominger <arominge@redhat.com>
             https://github.com/AlanCoding
2019-06-13 14:13:28 +00:00
AlanCoding
76ebcf914b upgrade pip and setuptools 2019-06-13 09:45:18 -04:00
mabashian
effe7151eb Adds missing semi-colon 2019-06-12 22:13:45 -04:00
mabashian
0023591bb0 Removes unnecessary variable 2019-06-12 22:13:45 -04:00
mabashian
9c50609776 Always show search bar in smart inventory host filter modal 2019-06-12 22:13:45 -04:00
softwarefactory-project-zuul[bot]
28cc08f215 Merge pull request #4033 from mabashian/bootstrap-whitelist-table
Whitelist table elements for use in bootstrap popovers

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-06-12 21:12:37 +00:00
mabashian
b83cef6ed7 Fixes links in various lists to maintain query params while navigating to edit form 2019-06-12 16:12:08 -04:00
mabashian
c729b698dd Whitelist table elements for use in bootstrap popovers 2019-06-12 15:32:08 -04:00
softwarefactory-project-zuul[bot]
e70c7ab458 Merge pull request #3967 from jbradberry/upgrade-drf-3.9.4
Upgrade DRF to 3.9.4

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-06-12 18:41:05 +00:00
Jeff Bradberry
aa548442ce Ensure the request is on the Metadata object
in order to correctly populate the default for the TOWER_URL_BASE
field, which is specific to the host and preferred scheme.
2019-06-12 14:12:59 -04:00
Jeff Bradberry
ab587e7e6c Update get_view_description to conform to changes in DRF's version
related encode/django-rest-framework#5605
2019-06-12 14:12:59 -04:00
Jeff Bradberry
3f5c018c8a Replace the load of the staticfiles template library with static
The staticfiles library will be going away before too long.
2019-06-12 14:12:59 -04:00
Jeff Bradberry
12e3d0aebf Remove rest_framework/base.html 2019-06-12 14:12:59 -04:00
Jeff Bradberry
ba4ad191fc Add the csrf token to window.drf
In the same manner as the current version of base.html from DRF.  This
is part 2 of the removal of base.html.
2019-06-12 14:12:59 -04:00
Jeff Bradberry
34d76422d6 Move the endpoint deprecation warning out of base.html
This is part 1 of the removal of awx/templates/rest_framework/base.html.
2019-06-12 14:12:59 -04:00
Jeff Bradberry
2a81643308 Refactor the SSO serializer fields to follow the DRF idioms more closely
and fix the tests to handle the newer nested validation checks properly.
2019-06-12 14:12:59 -04:00
Jeff Bradberry
76d4de24df Handle a change in the error message for BooleanField
related encode/django-rest-framework#5881
2019-06-12 14:12:59 -04:00
Jeff Bradberry
ed7a7e5f7b Support parse_requirements out of pip < 10 and >= 10 2019-06-12 14:12:59 -04:00
Jeff Bradberry
f94959d120 Remove the custom get_view_name function
It appeared to not be doing anything that we were making use of that
couldn't already be done, slightly differently, using DRF's built-in
one.
2019-06-12 14:12:59 -04:00
Jeff Bradberry
b5728fc548 Bump Django Rest Framework from 3.7.7 to 3.9.4 2019-06-12 14:12:59 -04:00
softwarefactory-project-zuul[bot]
4a19da650d Merge pull request #4056 from ansible/capitalization
updated e2e to work with new capitalization

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-06-12 17:36:29 +00:00
Daniel Sami
8db27611ca updated e2e to work with new capitalization 2019-06-12 13:05:11 -04:00
softwarefactory-project-zuul[bot]
7161f28d26 Merge pull request #4053 from AlanCoding/apps_management
Correctly manage current apps for the action_node addition

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-06-12 15:59:55 +00:00
softwarefactory-project-zuul[bot]
290c242221 Merge pull request #4054 from ansible/unpin-vault-plugin-container-image
unpin vault plugin development container image

Reviewed-by: awxbot
             https://github.com/awxbot
2019-06-12 15:00:23 +00:00
Jake McDermott
90fb7c6769 unpin vault plugin development container image 2019-06-12 10:23:01 -04:00
AlanCoding
f8c69aadcb Correctly manage current apps for the action_node addition 2019-06-12 10:13:45 -04:00
softwarefactory-project-zuul[bot]
c7b38bc9b9 Merge pull request #4050 from Spredzy/capitalize
User: Ensure First Name is not capitalized

Reviewed-by: Ryan Petrello
             https://github.com/ryanpetrello
2019-06-12 13:23:21 +00:00
Yanis Guenane
cc1ef50729 User: Ensure First Name is not capitalized
When creating a new user, the Last Name is not capitalized but First
Name is. Make the behavior being consistent.

Signed-off-by: Yanis Guenane <yguenane@redhat.com>
2019-06-12 14:40:44 +02:00
softwarefactory-project-zuul[bot]
42b3aa45c5 Merge pull request #4045 from ansible/quick-chrome75-workaround
Chrome75 workaround for e2e tests

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-06-12 11:52:09 +00:00
John Hill
1e91e4e531 set to json protocol 2019-06-11 15:51:02 -04:00
softwarefactory-project-zuul[bot]
b3979eb2b9 Merge pull request #4042 from shanemcd/cache-me-outside
Always consider devel tag as a cache source

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-06-11 18:34:25 +00:00
Shane McDonald
75ef30d21b Always consider devel tag as a cache source
When you do `make docker-compose-build` on your branch without any changes, this will use the devel tag as a cache source. This will speed things up if you havent made any local changes to the Makefile or requirements files.
2019-06-11 14:01:35 -04:00
softwarefactory-project-zuul[bot]
31b78cc00f Merge pull request #4036 from AlanCoding/event_query
Cache smart inventory memberships to avoid slow query

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-06-11 15:55:55 +00:00
softwarefactory-project-zuul[bot]
a510f9f2c7 Merge pull request #4038 from jakemcdermott/update-v1-cred-delete-prompt-queries
update credential deletion prompt queries

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-06-11 15:26:49 +00:00
AlanCoding
05aab5da4c cache smart inventory memberships to avoid slow query 2019-06-11 11:06:35 -04:00
softwarefactory-project-zuul[bot]
3054fbc61c Merge pull request #4040 from ryanpetrello/ldap-py3-bug
fix a py3 compat problem in an LDAP migration

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-06-11 14:47:12 +00:00
Ryan Petrello
be8a30b9d9 fix a py3 compat problem in an LDAP migration 2019-06-11 10:25:15 -04:00
Jake McDermott
d4301bd9bd update v1 credential deletion prompt queries 2019-06-11 09:45:15 -04:00
softwarefactory-project-zuul[bot]
adb768bed3 Merge pull request #3413 from ryanpetrello/bye-bye-v1
remove /api/v1 and deprecated credential fields

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-06-10 17:31:28 +00:00
softwarefactory-project-zuul[bot]
9253ab28c8 Merge pull request #4021 from ryanpetrello/dotted-vault-id
support vault IDs that include dot characters

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-06-06 18:34:24 +00:00
softwarefactory-project-zuul[bot]
31d4e8362e Merge pull request #4003 from keithjgrant/3574-survey-tab-required-fields
Don't allow Enter to proceed to next step if survey fields invalid

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-06-06 18:33:29 +00:00
Ryan Petrello
329b40fd69 support vault IDs that include dot characters
see: https://github.com/ansible/awx/issues/4009
2019-06-06 12:42:31 -04:00
Jake McDermott
6bc5c4da74 include credential in inventory update detail summary 2019-06-06 12:23:17 -04:00
Jake McDermott
51d7de296f remove dead code related to v1 kind field
Currently, the credentials list doesn't seem to be returning
any options data for 'kind' so this code wasn't being reached. In
the future api updates, we'll also be removing the 'kind' field from
credentials in general.
2019-06-06 12:23:11 -04:00
Jake McDermott
5987aafb82 update shared client code for api v2 2019-06-06 12:23:09 -04:00
Jake McDermott
7a0a2fb54c update adhoc command queries for api v2 2019-06-06 12:23:06 -04:00
Jake McDermott
2f57a1ea93 update inventory source credential lookup queries for api v2 2019-06-06 12:23:03 -04:00
Ryan Petrello
6da445f7c0 remove /api/v1 and deprecated credential fields 2019-06-06 12:23:00 -04:00
softwarefactory-project-zuul[bot]
176f8632e5 Merge pull request #4017 from ryanpetrello/swagger-action-node
make activity stream action_node deterministic for Swagger docs

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-06-06 16:13:17 +00:00
softwarefactory-project-zuul[bot]
958c192ff7 Merge pull request #4000 from rooftopcellist/order_job_counts
fix analytics job counts by ordering query

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-06-06 15:16:42 +00:00
softwarefactory-project-zuul[bot]
7e8990dff9 Merge pull request #4011 from shanemcd/devel
Clean up related to isolated env

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-06-06 14:48:49 +00:00
Christian Adams
a727de184b Fix analytics job counts by ordering query
- More Info: https://github.com/ansible/awx/issues/4014
2019-06-06 10:45:26 -04:00
Ryan Petrello
6d1ba411e6 make activity stream action_node deterministic for Swagger docs 2019-06-06 10:36:34 -04:00
Shane McDonald
b00979792e Clean up related to isolated env
- Made  make target names consistent with our other commands
- Found some unnecessary code, deleted it
2019-06-06 10:15:11 -04:00
softwarefactory-project-zuul[bot]
8be0b01c33 Merge pull request #4013 from ansible/docker-container-tag-updates
Updated docker tags in container setup

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-06-05 20:17:03 +00:00
Daniel Sami
62a3e0df98 Updated docker tags in container setup
Use env variable for selenium docker tag
2019-06-05 15:10:06 -04:00
softwarefactory-project-zuul[bot]
c7f49c1193 Merge pull request #3992 from marshmalien/isolated-badge
Add isolated badge to isolated instance groups

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-06-05 16:58:20 +00:00
softwarefactory-project-zuul[bot]
3fcf3b20c4 Merge pull request #4005 from shanemcd/sdist-builder-node
Update node in AWX installer sdist builder

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-06-05 11:48:30 +00:00
softwarefactory-project-zuul[bot]
5db3a8e7dc Merge pull request #4007 from git4anand/devel
Multiple installation of same dependency

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-06-05 11:39:46 +00:00
Anand kumar
3d7bd8579b Multiple installation of same dependency
python-pip is being installed multiple times.
2019-06-05 12:38:52 +05:30
Shane McDonald
99704af302 Fix some ansible warnings 2019-06-04 20:37:03 -04:00
Shane McDonald
a13b733191 Update node in sdist builder image 2019-06-04 20:36:39 -04:00
Keith Grant
4a7cd56e4a do allow Enter to proceed to next step if survey fields invalid 2019-06-04 16:08:23 -07:00
softwarefactory-project-zuul[bot]
839f49c6ed Merge pull request #3997 from ryanpetrello/libcloud
unpin apache-libcloud now that a version is out without PyCrypto

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-06-04 19:59:06 +00:00
Ryan Petrello
c6afd98500 unpin apache-libcloud now that a version is out without PyCrypto
see: https://github.com/ansible/awx/pull/3466
see: https://github.com/ansible/awx/issues/3822
2019-06-04 09:26:27 -04:00
softwarefactory-project-zuul[bot]
67fb898a9d Merge pull request #3993 from shanemcd/fix-isolated-dev-env
Fix isolated dev env

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-06-03 21:43:31 +00:00
Shane McDonald
f18f9ec0ef Bring isolated dev env current
There was quite a bit of drift between the main Dockerfile and this one since we last touched it. It’s heavier than it needs to be, but by using the main development image as the base here, we have less duplicated and outdated code overall.
2019-06-03 16:48:45 -04:00
Shane McDonald
ef22986aa0 Install latest tini from RPM 2019-06-03 16:03:51 -04:00
softwarefactory-project-zuul[bot]
1829017ad4 Merge pull request #3991 from shanemcd/downgrade-ptyprocess
Downgrade ptyprocess

Reviewed-by: awxbot
             https://github.com/awxbot
2019-06-03 18:19:15 +00:00
Shane McDonald
be3d095067 Downgrade ptyprocess
See 8be490695d (diff-522adf759addbd3b193c74ca85243f7d) for more info
2019-06-03 13:42:45 -04:00
softwarefactory-project-zuul[bot]
112b9e7381 Merge pull request #3984 from jpmens/patch-1
Emphasize a recent version of Docker

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-06-03 17:42:24 +00:00
softwarefactory-project-zuul[bot]
87bd3c2f93 Merge pull request #3987 from Spredzy/quote_password
Installer: quote password where it applies

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-06-03 17:38:49 +00:00
Marliana Lara
abb37299cb Add isolated badge to instance groups list 2019-06-03 13:20:23 -04:00
softwarefactory-project-zuul[bot]
724ca9cd57 Merge pull request #3979 from elyezer/resize-window-e2e
e2e resize window on some specific tests

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-06-03 14:21:43 +00:00
Elyézer Rezende
39fb0d1679 e2e resize window on some specific tests 2019-06-03 08:50:37 -04:00
Yanis Guenane
11630a8803 Installer: quote password where it applies
Prior to this change, password having shell interpretable character
would break the installer (e.g '&', '(', etc... )

This commits rely on the `quote` filter from ansible to ensure those
password are properly quoted where it applies

Fixes: https://github.com/ansible/awx/issues/3943
Signed-off-by: Yanis Guenane <yguenane@redhat.com>
2019-06-03 11:50:52 +02:00
JP Mens
a7b96d5aec Emphasize a recent version of Docker
I had 1.13 installed as part of Centos Extras and spent hours attempting to install AWX 4.0.0; the attempts all threw masses of permission denied errors.

Uninstalling that version and replacing with a current docker-ce then worked.
2019-06-01 16:29:10 +02:00
softwarefactory-project-zuul[bot]
d8a80f9f3e Merge pull request #3981 from jakemcdermott/test-debug-helpers
add some debug helpers for tests

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-31 16:19:20 +00:00
Jake McDermott
1dcb1eda7c add some debug helpers for tests
Signed-off-by: Jake McDermott <yo@jakemcdermott.me>
2019-05-31 11:51:36 -04:00
softwarefactory-project-zuul[bot]
f64e31735c Merge pull request #3980 from wenottingham/following-in-finns-footsteps
Remove some things that are not first-order dependencies from requirements.in

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-31 15:14:43 +00:00
softwarefactory-project-zuul[bot]
d7c33a7246 Merge pull request #3893 from AlanCoding/replace_job_origin
Replace JobOrigin model with ActivityStream.action_node field

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-31 11:59:09 +00:00
AlanCoding
fedd1cf22f Replace JobOrigin with ActivityStream.action_node 2019-05-31 07:10:07 -04:00
Bill Nottingham
12ff7a481c Remove some things that are not first-order dependencies.
This includes a few things where the version specifiers resolve properly now.
2019-05-30 22:42:37 -04:00
softwarefactory-project-zuul[bot]
9b5494a6cc Merge pull request #3976 from Spredzy/fix_quote
openshift: Ensure char in password are not interpreted

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-31 01:27:38 +00:00
Yanis Guenane
99296cf5f1 openshift: Ensure char in password are not interpreted
If password contains ';' (and potentially any shell interpretable chars)
it won't be interpreted properly as the openshift password.

Signed-off-by: Yanis Guenane <yguenane@redhat.com>
2019-05-31 03:00:13 +02:00
softwarefactory-project-zuul[bot]
e6b78292ec Merge pull request #3974 from ryanpetrello/downstream-i18n
Fix python3 fallout in translations management script

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-30 19:06:26 +00:00
Shane McDonald
49b54f2d60 Fix python3 fallout in translations management script 2019-05-30 14:41:03 -04:00
softwarefactory-project-zuul[bot]
0a256a98be Merge pull request #3965 from AlexSCorey/2866-JTForksToolTip
Updates text of tooltip for Forks in JT.

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-30 17:09:49 +00:00
softwarefactory-project-zuul[bot]
5756151568 Merge pull request #3921 from joseorpa/devel3004proxy
#3004 Add proxy support to postgresql, memcached and rabbitmq images

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-30 14:33:17 +00:00
softwarefactory-project-zuul[bot]
6606fd7461 Merge pull request #3930 from kialam/upgrade-node-v8-to-node-v10
Upgrade our UI packaging for NodeJS 10.

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-30 14:29:52 +00:00
Jose OrPa
84b6866875 #3004 Add proxy support to postgresql, memcached and rabbitmq images 2019-05-30 10:04:17 -04:00
Kia Lam
4b9024bcb0 Revert puppeteer upgrade to fix failing UI unit tests in Zuul. 2019-05-30 09:51:19 -04:00
softwarefactory-project-zuul[bot]
4b1fc7894d Merge pull request #3961 from rooftopcellist/rm_implicit
remove implicit grant type for OAuth 2 apps

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-29 20:27:11 +00:00
Alex Corey
a5a6fdf1d6 Updates text of tooltip for Forks in JT. 2019-05-29 14:27:16 -04:00
Kia Lam
15c699de7c Update dockerfile to build using node v10.15.0. 2019-05-29 13:00:06 -04:00
Kia Lam
515d4fe20f Update INSTALL.md with new node version. 2019-05-29 12:21:54 -04:00
Kia Lam
75380b9576 Upgrade our packaging for NodeJS 10.
- Update README.md
- Bump certain dependencies to fix vulnerabilities. Exclude dependency bumps that would cause breaking changes.
2019-05-29 12:21:54 -04:00
Christian Adams
cb279843d2 remove implicit grant type for OAuth 2 apps 2019-05-29 02:28:37 -04:00
softwarefactory-project-zuul[bot]
41f2b83ae2 Merge pull request #3947 from ryanpetrello/transient-queues
RFC: add the ability to disable RabbitMQ queue durability

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-28 20:15:26 +00:00
Ryan Petrello
40b1e89b67 add the ability to disable RabbitMQ queue durability 2019-05-28 15:49:32 -04:00
softwarefactory-project-zuul[bot]
8c56d1d3a7 Merge pull request #3932 from shanemcd/zuul_k8s
Changes to enable tests in k8s

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-28 19:34:58 +00:00
Shane McDonald
2f77c67a62 Changes to enable tests in k8s 2019-05-28 13:22:15 -04:00
softwarefactory-project-zuul[bot]
5a502f8709 Merge pull request #3925 from AlanCoding/gce_contract
Restore old use pattern for gce_ modules, update token_uri URL

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-28 17:03:32 +00:00
softwarefactory-project-zuul[bot]
873ff3de78 Merge pull request #3928 from wenottingham/waka-flocka
Use lockf, not flock.

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-28 16:29:23 +00:00
softwarefactory-project-zuul[bot]
eb3ef809e0 Merge pull request #3873 from shanemcd/faster-boots
Avoid DNS timeout in non-Docker for Mac installs

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-28 14:30:45 +00:00
softwarefactory-project-zuul[bot]
9c90694f12 Merge pull request #3604 from athenahealth/complete-ssl-support
Update SSL support for docker-compose install

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-28 13:51:43 +00:00
softwarefactory-project-zuul[bot]
ca3735ee73 Merge pull request #3941 from elyezer/window-size-e2e
Set window size globally on e2e

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-24 20:23:45 +00:00
softwarefactory-project-zuul[bot]
380f122456 Merge pull request #3804 from AlanCoding/more_optimizations
Apply optimizations via standard methods

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-24 19:38:46 +00:00
Elyézer Rezende
0aa8c7427d Set window size globally on e2e 2019-05-24 14:05:39 -04:00
AlanCoding
dbc65baa43 apply optimizations via standard method
This addresses the top-level resources in the v2
root view, focusing in order of priority,
reflecting use by the UI.

In several cases get_queryset logic from the view
is moved into the access class.

Most other cases involve adding a straightforward
select_related or prefetch_related entry.

All additional confirmed to be effective with the
django debug toolbar.
2019-05-24 13:09:55 -04:00
softwarefactory-project-zuul[bot]
1ce587025e Merge pull request #3948 from rooftopcellist/fix_analytics_log_msg
fix analytics logging msg

Reviewed-by: awxbot
             https://github.com/awxbot
2019-05-23 18:55:55 +00:00
Christian Adams
45458b3265 fix analytics logging msg 2019-05-23 14:26:03 -04:00
softwarefactory-project-zuul[bot]
874465a2d4 Merge pull request #3865 from chrismeyersfsu/fix-enabled_still_online
disabled instance does not mean offline instance

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-23 16:55:09 +00:00
softwarefactory-project-zuul[bot]
34c3aaee3d Merge pull request #3682 from AlanCoding/depgrades
General upgrade of dependencies

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-23 15:15:20 +00:00
softwarefactory-project-zuul[bot]
13ff5ffdf2 Merge pull request #3900 from AlanCoding/fewer_type_methods
Remove duplicated type methods and old Django logic

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-23 12:50:35 +00:00
softwarefactory-project-zuul[bot]
a606fdc958 Merge pull request #3933 from AlanCoding/schedule_no_op
More precise handling of schedule computed fields no-ops

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-23 12:39:58 +00:00
AlanCoding
b11995e638 Remove duplicated type methods and old Django logic 2019-05-23 08:24:24 -04:00
AlanCoding
40f9d0b512 More precise handling of schedule computed fields no-ops
Do not set a next_run value for disabled schedules
Bail if no fields are changed
Do not update related template if its fields did not change

Change call pattern to schedule.update_computed_fields()
in doing so, fix bug where template does not pick up schedule
  due to schedules next_run not being saved

Handle the case (also a bug) where template was not updated
  when schedule was deleted
2019-05-23 08:13:37 -04:00
softwarefactory-project-zuul[bot]
9a1b205e06 Merge pull request #3944 from AlanCoding/move_to_filter
Move dynamic log level logic to filter

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-22 23:45:36 +00:00
softwarefactory-project-zuul[bot]
98c923a715 Merge pull request #3856 from rooftopcellist/cleanup_sessions_refactor
Cleanup Sessions & Tokens System Jobs

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-22 20:50:48 +00:00
softwarefactory-project-zuul[bot]
1d328134fd Merge pull request #3931 from wenottingham/be-bold
Update notification radio button styling to match schedule radiobuttons.

Reviewed-by: Bill Nottingham
             https://github.com/wenottingham
2019-05-22 18:58:43 +00:00
AlanCoding
d3f047d731 Move dynamic log level logic to filter 2019-05-22 14:00:39 -04:00
Christian Adams
8ca0c1b992 Add clearsessions and cleartokens system jobs
* add system job for gathering insights analytics
* enforce schedule enablement with analytics setting
* remove celery beat analytics job
* keep analytics schedule & setting enablement in sync in API
* handles updating schedules for multiple sys job schedules
* add analytics setting & schedule test
* rm ui modal from collection sys job
2019-05-22 13:11:48 -04:00
softwarefactory-project-zuul[bot]
4a711ec2dc Merge pull request #3914 from itdependsnetworks/var_port_numbers
Update the ports for memcache and rabbitmq to be variablized

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-22 14:08:41 +00:00
softwarefactory-project-zuul[bot]
75fe801efb Merge pull request #3913 from marshmalien/3886-instance-group-breadcrumb-label
Dynamically show instance group name in breadcrumb

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-21 14:28:55 +00:00
Marliana Lara
56df930b99 Sanitize instance group breadcrumb name 2019-05-21 09:55:35 -04:00
Bill Nottingham
f48713f4ae Use lockf, not flock.
This performs more reliably on certain filesystems in Linux.
2019-05-20 16:56:16 -04:00
Bill Nottingham
bb009f0d12 Update notification radio button styling to match schedule radiobuttons. 2019-05-20 16:46:10 -04:00
softwarefactory-project-zuul[bot]
dc1bf3ef07 Merge pull request #3910 from AlanCoding/no_user_get
Avoid unnecessary user get expiring session memberships

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-20 17:20:36 +00:00
softwarefactory-project-zuul[bot]
9d4cfa7400 Merge pull request #3890 from AlanCoding/fix_debug_toolbar
Fix Django debug toolbar after its upgrade

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-20 16:41:13 +00:00
AlanCoding
06be3a29b9 fix Django debug toolbar after its upgrade 2019-05-20 12:17:57 -04:00
softwarefactory-project-zuul[bot]
2addf20907 Merge pull request #3918 from jbradberry/middleware-ordering
Fix the middleware ordering

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-20 15:43:06 +00:00
softwarefactory-project-zuul[bot]
29bbecb6bf Merge pull request #3820 from AlanCoding/default_ordering
Resolve default ordering warnings from tests

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-20 15:31:36 +00:00
AlanCoding
f4c18843a3 Resolve default ordering warnings from tests 2019-05-20 10:58:36 -04:00
AlanCoding
bda838f723 Add note for docutils, unpin runner dependency 2019-05-20 10:24:43 -04:00
AlanCoding
2bec5ddb41 General upgrade of dependencies
Update licenses for new versions after dependency upgrades

pin pycurl to version that does not break on install

implement new workflow for py2/3 requirements management

require twisted tls extras, resolve service-identity version

Upgrade celery to resolve importlib DeprecationWarning

use flags to resolve the unsafe and cache problems
2019-05-20 10:24:39 -04:00
AlanCoding
74643520c7 GCE contract update, do not use token_uri in job runs 2019-05-20 09:27:45 -04:00
Jeff Bradberry
44907b33dc Fix the middleware ordering 2019-05-17 14:40:33 -04:00
softwarefactory-project-zuul[bot]
f174902bb2 Merge pull request #3874 from jbradberry/expose-role-type
Expose the content type associated with roles for ActivityStream objects in the API

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-17 18:06:38 +00:00
AlanCoding
1223148116 Avoid unnecessary user get expiring session memberships 2019-05-17 14:05:17 -04:00
softwarefactory-project-zuul[bot]
ab1e45d6c4 Merge pull request #3875 from rooftopcellist/swifty_operation
reorder migrations to avoid session error

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-17 15:41:26 +00:00
softwarefactory-project-zuul[bot]
bd50e5d6a8 Merge pull request #3896 from rooftopcellist/collection_no_reset
collect events based on last ship, not collection

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-17 15:32:48 +00:00
itdependsnetworks
f085b828e4 Update the ports for memcache and rabbitmq to be variablized 2019-05-16 19:59:57 -04:00
Marliana Lara
e95339ba6e Add instance group name to breadcrumb 2019-05-16 16:46:53 -04:00
softwarefactory-project-zuul[bot]
d353daebc5 Merge pull request #3911 from rooftopcellist/rm_extra_error
rm extra collection error line

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-16 20:11:41 +00:00
softwarefactory-project-zuul[bot]
6681cd918c Merge pull request #3906 from heavenly999/devel
Increasing requests-credssp ver to 1.0.2

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-16 20:00:07 +00:00
Christian Adams
2b327935de reorder migrations to avoid session error 2019-05-16 15:13:47 -04:00
Christian Adams
0c4925afe8 rm extra collection error line 2019-05-16 14:48:41 -04:00
softwarefactory-project-zuul[bot]
e0062484d0 Merge pull request #3908 from rverchere/fr_fix
Correct fr translation for Cancel message

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-16 18:46:19 +00:00
softwarefactory-project-zuul[bot]
faa353521a Merge pull request #3909 from marshmalien/3884-dropdown-alignment
Fix styling bug in Sort Dropdown

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-16 18:03:07 +00:00
softwarefactory-project-zuul[bot]
ff7e244a84 Merge pull request #3905 from NickBusey/patch-1
Update INSTALL.md

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-16 17:02:07 +00:00
Marliana Lara
17f71600df Add styles to grow dropdown width based on content 2019-05-16 12:23:47 -04:00
Rémi VERCHERE
9120a69006 Correct fr translation for Cancel message 2019-05-16 17:42:13 +02:00
softwarefactory-project-zuul[bot]
d9965cfe7e Merge pull request #3882 from ryanpetrello/debug-log
add the ability to toggle DEBUG logging on dynamically

Reviewed-by: Shane McDonald <me@shanemcd.com>
             https://github.com/shanemcd
2019-05-16 15:12:01 +00:00
Adam Nagy
c38ee06642 Increasing requests-credssp ver to 1.0.2
Signed-off-by: Adam Nagy <anagy@netsuite.com>
2019-05-16 15:56:35 +02:00
Ryan Petrello
b1d75327e3 add the ability to toggle DEBUG logging on dynamically 2019-05-16 07:58:31 -04:00
softwarefactory-project-zuul[bot]
8981c7d59a Merge pull request #3904 from wenottingham/new-year-new-you
Bump date.

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-16 03:35:31 +00:00
Nick Busey
b10f06201d Update INSTALL.md 2019-05-15 16:34:27 -06:00
Bill Nottingham
681fe4865c Some more dates. 2019-05-15 17:01:33 -04:00
Bill Nottingham
faae55d085 Bump date. 2019-05-15 16:58:59 -04:00
softwarefactory-project-zuul[bot]
efddd9f679 Merge pull request #3854 from beeankha/add_debug_toolbar
Update Custom Middleware to New Style

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-15 18:38:24 +00:00
beeankha
6fb173da8a Remove redundant methods from SocialAuthMiddleware class 2019-05-15 12:54:00 -04:00
beeankha
64e8b76a10 Remove redundant middleware 2019-05-15 10:16:20 -04:00
Christian Adams
d0e160a037 collect events based on last ship, not collection 2019-05-14 17:14:44 -04:00
beeankha
20e5d8200e Subclass more middlware with deprecation mixin 2019-05-14 13:37:34 -04:00
beeankha
a6d3c0fd32 Remove redundant code and update URLModificationMiddleware 2019-05-14 13:37:34 -04:00
beeankha
318e0631b7 Add super() call 2019-05-14 13:37:34 -04:00
beeankha
da4153d653 Add mixin to ActivityStreamMiddleware class 2019-05-14 13:37:34 -04:00
beeankha
26e9dd307e Fix missing argument error 2019-05-14 13:37:34 -04:00
beeankha
6a2d59963f Update wsgi file to be compatible with new style of middleware 2019-05-14 13:37:34 -04:00
beeankha
68800d0e8e Make custom middleware use new style vs old 2019-05-14 13:37:34 -04:00
beeankha
97dc77ea63 Add debug_toolbar.middleware.DebugToolbarMiddleware to MIDDLEWARE classes in development.py 2019-05-14 13:37:34 -04:00
softwarefactory-project-zuul[bot]
6bc1856658 Merge pull request #3861 from beeankha/awx_doc_edits
Update/Edit AWX Docs

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-13 20:41:14 +00:00
beeankha
9ea3ec24ca Address comments/suggested changes 2019-05-13 15:46:24 -04:00
beeankha
7466873f69 Make more edits/grammatical changes 2019-05-13 15:27:54 -04:00
beeankha
07fa533b6f Make edits to Notification doc 2019-05-13 11:47:04 -04:00
chris meyers
8aa28092ff disabled instance does not mean offline instance
* Disabling an instance is used to stop and instance from being the
target of new jobs to run.
* The instance should still perform it's heartbeat so that it isn't
considered offline.
* If the instance was allowed to go offline on an openshift cluster it
would be deleted from the database.
2019-05-13 11:44:47 -04:00
Jeff Bradberry
3579584ffc i18n 2019-05-13 11:20:15 -04:00
Jeff Bradberry
05cae23180 Add help text to the ActivityStreamSerializer m2m relationship fields 2019-05-13 11:10:33 -04:00
Jeff Bradberry
d6e89092d3 Instead of exposing Role.content_type, create a new serializer field
called `object_type`, which is constructed based on manipulating the
string value of ActivityStream.object_relationship_type.  Since that
field does have the full class name, this manipulation should match
the manipulation that is done to construct the values of object1 and
object2 when ActivityStream is created.
2019-05-10 17:31:33 -04:00
beeankha
fe344038b5 Edit Clustering doc 2019-05-10 14:54:20 -04:00
Jeff Bradberry
250484339b Expose the role's content_type field and reorder the objects in the UI
when the resource was stored on ActivityStream as object1 instead of object2.

related #3841
2019-05-10 14:15:05 -04:00
Shane McDonald
5ca0cdb124 Avoid DNS timeout in non-Docker for Mac installs
Shaves 20 seconds off of rebooting the dev environment on Linux.
2019-05-10 12:31:08 -04:00
softwarefactory-project-zuul[bot]
0d3f1f4ac2 Merge pull request #3871 from ryanpetrello/devel
merge in downstream changes

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-10 16:21:39 +00:00
softwarefactory-project-zuul[bot]
18c69fa391 Merge pull request #3870 from shanemcd/gimme-root
Ensure that a root user is always present in development environment

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-10 14:57:58 +00:00
softwarefactory-project-zuul[bot]
6d6eae571e Merge pull request #3868 from jakemcdermott/ig-order-docstring
add docstring and test for instance group service

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-10 14:54:52 +00:00
Shane McDonald
7d8a910be7 Improve dev environment init process
This ensures that /etc/passwd is always written, regardless of how the container starts.
2019-05-10 10:14:51 -04:00
Jake McDermott
df04660cdd add test for instance groups service 2019-05-10 10:05:05 -04:00
softwarefactory-project-zuul[bot]
9d1ed837f9 Merge pull request #3869 from AlanCoding/dev_supervisor_fix
Fix bug in dev supervisor reparenting processes

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-10 13:45:30 +00:00
Ryan Petrello
be4705ef8b Merge branch 'downstream' into devel 2019-05-10 09:18:07 -04:00
Shane McDonald
015234287c Ensure that a root user is always present in development environment
@AlanCoding was seeing errors in the development container when trying to run some commands as root. This fixes that.
2019-05-10 09:16:28 -04:00
AlanCoding
677a8b34ba Fix bug in dev supervisor reparenting processes 2019-05-10 07:46:53 -04:00
Jake McDermott
232c706b75 add docstring for instance group service 2019-05-10 07:16:50 -04:00
softwarefactory-project-zuul[bot]
8725d3e539 Merge pull request #3842 from ryanpetrello/instance-group-order
enforce a stable list order when attaching/detaching instance groups

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-09 21:04:29 +00:00
softwarefactory-project-zuul[bot]
e7290e6452 Merge pull request #3864 from rooftopcellist/more_verbose_instance_msg
more verbose list_instance messaging

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-09 19:31:24 +00:00
Christian Adams
21105b836e more verbose list_instance messaging 2019-05-09 15:08:56 -04:00
softwarefactory-project-zuul[bot]
99dc84c275 Merge pull request #3863 from ryanpetrello/makefile-supervisor
use make targets for dev supervisor commands

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-09 18:07:39 +00:00
Ryan Petrello
adfdfcdd0a use make targets for dev supervisor commands 2019-05-09 13:45:49 -04:00
softwarefactory-project-zuul[bot]
6feb58f76d Merge pull request #3853 from AlanCoding/variable_data_permission
Put variable data permission in its own class

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-09 16:48:11 +00:00
softwarefactory-project-zuul[bot]
2910a9dfff Merge pull request #3858 from shanemcd/unsquash
Remove --squash option from awx-devel-build

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-09 14:19:59 +00:00
beeankha
371966613f Update AWX docs 2019-05-09 10:03:57 -04:00
softwarefactory-project-zuul[bot]
91968a09c8 Merge pull request #3849 from AlanCoding/password_madness
Reduce passing around of passwords dictionary

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-09 13:34:43 +00:00
softwarefactory-project-zuul[bot]
190098bbd5 Merge pull request #3848 from AlanCoding/standard_injector
Move outlier gce cache logic to standard place

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-09 12:56:11 +00:00
AlanCoding
2585c5030b move outlier gce cache logic to standard place 2019-05-09 08:24:57 -04:00
Shane McDonald
22858f0044 Remove --squash option from awx-devel-build
This was intended to save us CI time if / when we switch away from static nodes for Zuul. After merging this in I discovered that this option does not work with `docker build --cache-from`, which will be more benefical in terms of speed improvements anyway.
2019-05-08 22:22:38 -04:00
AlanCoding
7e6a73f892 fix bug with null credential 2019-05-08 21:18:34 -04:00
AlanCoding
1874e8bb4c Reduce passing around of passwords dictionary 2019-05-08 21:18:34 -04:00
softwarefactory-project-zuul[bot]
83c286580b Merge pull request #3852 from shanemcd/dockerfiled
Refactor Dockerfiles

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-09 01:08:49 +00:00
Shane McDonald
0be8fe521a Refactor Dockerfiles
This commit does a few things:

- Add the `--squash` option to the `awx-devel-build` make target. This reduces the resulting image size from 2.12 GB to 1.37 GB. I think we can get this down even more by inspecting the image contents.
- Reorganize commands so that the cache expires less often. Before this commit, any changes to the Makefile would essentially cause the entire image to rebuild.
- Break yum dependencies up into multiple lines. This makes it easier to see what changes in a diff.
- Use `n` to install our required version of node (rather `curl node | bash`). I’ve found this to be easier to maintain / more portable when working with other Dockerfiles.
- General organizational changes to make things easier to parse visually.
2019-05-08 20:36:41 -04:00
softwarefactory-project-zuul[bot]
186ec88581 Merge pull request #3839 from zicklam/webhook_disable_ssl_verify
Add "Disable SSL Verification" checkbox to webhook notification

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-08 19:44:11 +00:00
softwarefactory-project-zuul[bot]
6407ab58ff Merge pull request #3855 from ryanpetrello/improved-dev-code-reload
remove honcho in the dev environment (just use foregrounded supervisor)

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-08 19:25:57 +00:00
Michael Abashian
466e965047 Merge pull request #3549 from mabashian/host_filter_or
Fixes basic host filter searches
2019-05-08 15:16:43 -04:00
mabashian
e1de0a528d Adds some test coverage for host_filter queries 2019-05-08 14:59:04 -04:00
Ryan Petrello
766a5c0c3f remove honcho in the dev environment (just use foregrounded supervisor)
using supervisor gives us the ability to restart entire processes on
code change (like the dispatcher and callback receiver)
2019-05-08 14:44:07 -04:00
AlanCoding
231abf865b put variable data permission in its own class 2019-05-08 13:43:13 -04:00
softwarefactory-project-zuul[bot]
70972f7ea1 Merge pull request #3850 from jbradberry/reload-dispatcher-on-code-change
Make the uwsgi autoreloader reload the dispatcher every time

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-08 16:11:04 +00:00
Ryan Petrello
993b0a889d add API tests for InstanceGroup order persistence 2019-05-08 11:56:30 -04:00
Jake McDermott
205f2c33c1 wait for instance group requests on template save 2019-05-08 11:56:19 -04:00
Jake McDermott
1e77053bbf handle re-ordering of instance groups 2019-05-08 11:56:05 -04:00
Jeff Bradberry
ae25717700 Make the uwsgi autoreloader reload the dispatcher every time
not just the first time uwsgi is brought up.

related #3846
2019-05-08 11:15:54 -04:00
mabashian
11244f85a4 Adds missing semicolon 2019-05-08 10:05:02 -04:00
softwarefactory-project-zuul[bot]
e05c6e67b6 Merge pull request #3837 from saito-hideki/issue/2891
Add credential info in expanded list view of inventory update

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-08 13:37:34 +00:00
zicklam
42f30e72b5 False is not false 2019-05-08 08:35:26 +02:00
zicklam
0fb3851a2b webhook_notification set default for var 'disable_ssl_verification' 2019-05-08 08:21:04 +02:00
Ryan Petrello
e4a50f3595 enforce a stable list order when attaching/detaching instance groups 2019-05-07 14:53:00 -04:00
mabashian
f524c94bad Fixes basic host filter searches 2019-05-07 14:43:53 -04:00
softwarefactory-project-zuul[bot]
c7fe840868 Merge pull request #3810 from vismay-golwala/restrict_copy_jt
Raise meaningful error when permission denied to copy JT

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-07 13:52:04 +00:00
Vismay Golwala
0154d80f19 Raise meaningful error when permission denied to copy JT
When a user doesn't have access to all the credentials of a job
template, they cannot copy the JT. However, currently we raise a
default `PermissionDenied`, which doesn't give the user insight
into what's wrong. So, this PR just adds a custom message indicating
that access to credentials is missing.

Signed-off-by: Vismay Golwala <vgolwala@redhat.com>
2019-05-06 15:02:36 -04:00
zicklam
08d60d0b78 Update test_notification Template for webhooks
- rename webhook_no_verify_ssl to disable_ssl_verification
2019-05-06 19:20:00 +02:00
softwarefactory-project-zuul[bot]
6908558acd Merge pull request #3833 from ryanpetrello/iso-mem-cpu
properly record Instance.cpu and Instance.memory for isolated nodes

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-06 16:14:18 +00:00
softwarefactory-project-zuul[bot]
4dc7178f3c Merge pull request #3785 from AlanCoding/no_output_replacements
Remove unused output_replacements logic

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-06 16:11:51 +00:00
Michael Abashian
b85cc716a4 Merge pull request #3542 from mabashian/3540-groups-layout
Fix display bug on host groups list
2019-05-06 11:16:30 -04:00
Hideki Saito
418521f4a3 Add credential info in expanded list view of inventory update
- Fixed issue #2891

Signed-off-by: Hideki Saito <saito@fgrep.org>
2019-05-06 13:28:47 +00:00
zicklam
5b8fba58e8 Add "Disable SSL Verification" checkbox to webhook notification
This commit will add a checkbox which will disable SSL verification on
the generic webhook notification type. This is required when using
self-signed certificates.
2019-05-06 13:12:41 +02:00
Michael Abashian
842d48810c Merge pull request #3545 from elyezer/fix-joblist-sorting-options
Fix job list sorting options grouping
2019-05-03 19:50:13 -04:00
Elyézer Rezende
70e513a3cf Fix job list sorting options grouping
Ensure that the sorting options are grouped just like every other
sorting options list.
2019-05-03 16:20:01 -04:00
Ryan Petrello
c6c14d4fb9 properly record Instance.cpu and Instance.memory for isolated nodes 2019-05-03 15:30:41 -04:00
Ryan Petrello
a874ed0424 Merge pull request #3543 from AlanCoding/migration_schmigration
Fix server error for in-flight migration
2019-05-03 11:59:02 -04:00
softwarefactory-project-zuul[bot]
cf6a103207 Merge pull request #3818 from ryanpetrello/devel
merge in downstream changes

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-03 14:36:56 +00:00
AlanCoding
d2e67aea19 Fix server error for in-flight migration 2019-05-03 10:20:31 -04:00
softwarefactory-project-zuul[bot]
e5cf5be18d Merge pull request #3815 from pmoravec/awx-pmoravec-tower-sos-plugin-improvements
sosreport plugin improvements

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-03 12:47:39 +00:00
softwarefactory-project-zuul[bot]
f26ae8ef13 Merge pull request #3714 from vismay-golwala/org_member_count
Show only member users for organization

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-03 12:36:29 +00:00
softwarefactory-project-zuul[bot]
755c3e89e2 Merge pull request #3801 from vismay-golwala/copy_credential
Copy credential bug - add owner too

Reviewed-by: Ryan Petrello
             https://github.com/ryanpetrello
2019-05-03 12:31:59 +00:00
softwarefactory-project-zuul[bot]
2800e89fd2 Merge pull request #3783 from AlanCoding/passwords_and_relaunching
Allow relaunching other user jobs with public vars

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-03 12:03:26 +00:00
Ryan Petrello
8d75fc5f56 Merge branch 'downstream' into devel 2019-05-03 07:58:25 -04:00
softwarefactory-project-zuul[bot]
17d2efde95 Merge pull request #3789 from AlanCoding/rm_fields
Remove job ask_ fields that reference JT

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-03 11:42:19 +00:00
Pavel Moravec
8909a8a8e4 sosreport plugin to stick on sos >= 3.0 API only
As sosreport 3.0 was released 5 years ago, older sos versions can
be ignored / not further supported.

Signed-off-by: Pavel Moravec <pmoravec@redhat.com>
2019-05-03 10:38:34 +02:00
mabashian
90a86f53ba Fix display bug on host groups list 2019-05-02 17:13:17 -04:00
Michael Abashian
4be05f1bf6 Merge pull request #3535 from mabashian/job-events-collapse
Fix inert event expand/collapse on refreshed job results
2019-05-02 16:41:13 -04:00
Pavel Moravec
29822ee140 don't collect data from other sosreport plugins
Tower plugin shouldn't collect data that other sosreport plugins collect.

Further, few code optimizations in calling sos API are made.

Signed-off-by: Pavel Moravec <pmoravec@redhat.com>
2019-05-02 21:45:09 +02:00
Ryan Petrello
7fc13b8bb5 Merge pull request #3538 from ryanpetrello/valid-netloc
require a valid netloc for Credential Type inputs w/ format=url
2019-05-02 15:08:48 -04:00
Ryan Petrello
e560dccd36 require a valid netloc for Credential Type inputs w/ format=url 2019-05-02 14:49:02 -04:00
Ryan Petrello
de56e20f11 Merge pull request #3536 from ryanpetrello/urllib3-1.24.3
pin urllib3 to 1.24.3 to address CVE-2019-9740
2019-05-02 14:27:10 -04:00
Michael Abashian
d4cc595630 Merge pull request #3510 from mabashian/3506-sort-more-fields
Add more sort options to new lists
2019-05-02 14:17:36 -04:00
mabashian
b754e0dbba Fix template sorting by project 2019-05-02 13:55:20 -04:00
softwarefactory-project-zuul[bot]
72fe6e400e Merge pull request #3802 from AlanCoding/playbook_search
Include playbook as a default search field

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-02 16:54:21 +00:00
Ryan Petrello
6bb9902588 pin urllib3 to 1.24.3 to address CVE-2019-9740 2019-05-02 11:59:01 -04:00
mabashian
04d22a930d Fix inert event expand/collapse on refreshed job results 2019-05-02 10:43:56 -04:00
softwarefactory-project-zuul[bot]
fe3a2d1a4e Merge pull request #3803 from ryanpetrello/matter-of-fact
remove references to the (now defunct) fact receiver

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-02 13:34:12 +00:00
Ryan Petrello
50f9c70afd remove references to the (now defunct) fact receiver 2019-05-01 23:48:05 -04:00
softwarefactory-project-zuul[bot]
5ab7f888f1 Merge pull request #3800 from AlanCoding/event_children_list
Populate event children list via parent_uuid

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-02 00:37:03 +00:00
AlanCoding
7d692d08f9 include playbook as a default search field 2019-05-01 16:15:00 -04:00
softwarefactory-project-zuul[bot]
f3c023a11f Merge pull request #3765 from beeankha/task_doc_update
Add Task Overview Details to tasks.md

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-01 20:02:15 +00:00
beeankha
a87c87b7c9 Add more runner-related details 2019-05-01 15:40:15 -04:00
AlanCoding
5f1aeb0f4e remove ask_vars from job_start endpoint 2019-05-01 13:39:25 -04:00
Vismay Golwala
cc001c9892 Copy credential bug - add owner too
While copying a credentials, we should set the new owner to person
who copies it. There was a bug while trying to do so and this PR
tries to address it.

Link: #3750

Signed-off-by: Vismay Golwala <vgolwala@redhat.com>
2019-05-01 13:15:16 -04:00
softwarefactory-project-zuul[bot]
a68ab19e16 Merge pull request #3791 from AlanCoding/clean_up_test_tasks
Make test_tasks.py durable to changing Ansible versions

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-05-01 17:14:11 +00:00
AlanCoding
530a7ba51d populate event children list via parent_uuid 2019-05-01 11:55:53 -04:00
beeankha
383fe278f5 Address a few more comments 2019-05-01 11:35:26 -04:00
Ryan Petrello
c41da766fb Merge pull request #3530 from ryanpetrello/you're-not-by-real-dad
fix a bug that caused parent_uuid not to be persisted
2019-05-01 10:21:35 -04:00
Ryan Petrello
46795cc71e fix a bug that caused parent_uuid not to be persisted 2019-05-01 10:13:59 -04:00
Ryan Petrello
4fac608890 Merge pull request #3528 from jakemcdermott/fix-3507
require url scheme for external credential type url inputs
2019-05-01 09:32:42 -04:00
Jake McDermott
84b21620b2 raise url string parsing error as validation error 2019-05-01 09:17:52 -04:00
mabashian
eaaad89a8c Removed inventory sorting from jobs list. Changed Last Run to Last Used in project sort filters. 2019-05-01 09:17:15 -04:00
Michael Abashian
6f309bd2d2 Merge pull request #3526 from jakemcdermott/fix-3524
use test callback from test form save hooks
2019-05-01 08:57:29 -04:00
Ryan Petrello
6e00038d35 Merge pull request #3516 from jakemcdermott/fix-3511
hide dashboard tips on graph teardown event
2019-04-30 23:01:25 -04:00
Ryan Petrello
ad4e413a36 Merge pull request #3515 from jakemcdermott/fix-3513
fix error on template prompt initialization when credential has prompt
2019-04-30 15:47:39 -04:00
beeankha
27ca5e1fd5 Add more details for several of the listed tasks 2019-04-30 15:32:52 -04:00
AlanCoding
4191b21052 make test_tasks.py durable to changing Ansible versions 2019-04-30 14:00:51 -04:00
Jake McDermott
9737ab620c require url scheme for credential type url inputs
This adds a url formatting type for credential input string fields
The validator for this formatting type will throw an error if the
provided url string doesn't have a url schema.
2019-04-30 13:41:07 -04:00
AlanCoding
81f0662161 remove job ask_ fields that reference JT 2019-04-30 13:35:58 -04:00
Jake McDermott
d1dc6007fd use test callback from test form save hooks
Enter key and other form submits are handled generically from the built-in
save hooks on the form controller. This adds implementations for those hooks
on the plugin test forms to make sure the expected handlers are always called.
2019-04-30 12:31:30 -04:00
Ryan Petrello
52276ebbab Merge pull request #3523 from wenottingham/failure-event-handling-is-an-option
Use AWX python interpreter for failure-event-handler.
2019-04-30 10:41:54 -04:00
AlanCoding
d204f12184 remove unneeded update_model call 2019-04-30 10:15:08 -04:00
AlanCoding
42dd3c5cf5 Remove unused output_replacements logic 2019-04-30 10:11:03 -04:00
AlanCoding
5720601a2e allow relaunching other user jobs with public vars 2019-04-30 08:07:45 -04:00
Bill Nottingham
5d1346b956 Use AWX python interpreter for failure-event-handler.
python3 isn't in the normal path when using SCLs.
2019-04-29 20:32:50 -04:00
Vismay Golwala
30d0130e79 Show only member users for organization
Signed-off-by: Vismay Golwala <vgolwala@redhat.com>
2019-04-29 14:20:57 -04:00
beeankha
4baecef866 Clean up references to Tower 2019-04-29 10:51:25 -04:00
softwarefactory-project-zuul[bot]
aaeb2d6fb9 Merge pull request #3740 from wenottingham/lets-re-compose-ourselves
Sync docker-compose dockerfile with the actual production image build.

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-26 19:05:43 +00:00
softwarefactory-project-zuul[bot]
c707e60bde Merge pull request #3664 from vismay-golwala/pagination_cap_api
Cap page_size in pagination urls

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-26 17:55:51 +00:00
mabashian
b7e26b3507 Fix linting error 2019-04-26 13:15:37 -04:00
softwarefactory-project-zuul[bot]
44fcf2e0e4 Merge pull request #3766 from ryanpetrello/devel
merge in downstream changes

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-26 17:09:48 +00:00
mabashian
3cd8d4f7e6 Adds more sort options to lists 2019-04-26 11:52:07 -04:00
softwarefactory-project-zuul[bot]
ddde669083 Merge pull request #3767 from ryanpetrello/cors-lite
document CORS middleware

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-26 15:32:35 +00:00
Jake McDermott
36e384e8ab hide dashboard tips on graph teardown event 2019-04-26 11:18:39 -04:00
Ryan Petrello
586c3e4583 document CORS middleware 2019-04-26 11:00:33 -04:00
Jake McDermott
14045c1017 fix undefined prop in template credential prompt initialization
The prompt initialization code had a reference to an undefined
prop. This updates the code to use the expected prop.
2019-04-26 10:34:00 -04:00
Ryan Petrello
d11dfd0a2b Merge branch 'downstream' into devel 2019-04-26 08:02:04 -04:00
beeankha
cb7914dfa4 Make edits per first round of reviews. 2019-04-25 17:16:39 -04:00
beeankha
639e01e884 Remove unnecessary line 2019-04-25 16:26:58 -04:00
beeankha
e4d6d51cf5 Add further details/overview of AWX and Tower tasks into documentation 2019-04-25 16:14:49 -04:00
mabashian
7f452ee8d1 Add more sort options to new lists 2019-04-25 15:32:25 -04:00
Ryan Petrello
27d74528c0 Merge pull request #3494 from rooftopcellist/analytics_csv
analytics table copies formatted as csv now
2019-04-25 14:48:59 -04:00
Ryan Petrello
0ada1e965f Merge pull request #3508 from ryanpetrello/runner-134
pin runner 1.3.4
2019-04-25 10:16:05 -04:00
Ryan Petrello
2808a852eb pin runner 1.3.4 2019-04-25 09:49:03 -04:00
Ryan Petrello
80393e9194 Merge pull request #3504 from wenottingham/im-not-going-to-be-ignored-dan
Kill the rabbitmq sos collection from our plugin.
2019-04-24 14:51:01 -04:00
Bill Nottingham
aa52e41c02 Kill the rabbitmq sos collection from our plugin.
It exists in upstream sosreport, and this can cause conflicts.
2019-04-24 14:15:17 -04:00
Ryan Petrello
d7cdec37f2 Merge pull request #3502 from ryanpetrello/metrics-license-counts
fix a few issues with license counts in /api/v2/metrics/
2019-04-24 12:37:03 -04:00
softwarefactory-project-zuul[bot]
f60857013e Merge pull request #3754 from AlanCoding/gce_env_var
Simplify gce inventory plugin injector

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-24 16:25:09 +00:00
Ryan Petrello
c9424f9af8 fix a few issues with license counts in /api/v2/metrics/
- switched these to gauges so people can track them over time
- fixed a typo that caused `free_instances` to always be zero
2019-04-24 12:23:07 -04:00
Ryan Petrello
66f883befe Merge pull request #3485 from wenottingham/aaaaaaaaaaaaaaaaaaaaaaazure
Update Azure requirements for Ansible stable-2.8 branch.
2019-04-24 12:05:58 -04:00
softwarefactory-project-zuul[bot]
fad0274373 Merge pull request #3686 from vismay-golwala/instance_group_delete
[WIP] Disallow deleting controller or isolated instance groups

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-24 15:19:19 +00:00
Michael Abashian
8c715fc6e1 Merge pull request #3497 from dsesami/e2e_backport_3.5.0
E2E testing: compatibility edit to work with diff versions of node
2019-04-24 10:03:53 -04:00
Daniel Sami
870ebb4b43 fix to work with diff versions of node 2019-04-24 09:44:41 -04:00
softwarefactory-project-zuul[bot]
e78ef82385 Merge pull request #3758 from ryanpetrello/dot-files
clean up some old dotfiles

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-24 13:06:20 +00:00
Matthew Jones
1eb565543c Merge pull request #3495 from ansible/markerfiletest
Adding a markerfile for e2e runs
2019-04-24 09:06:03 -04:00
John Hill
082a819160 removing hyphen 2019-04-24 08:22:07 -04:00
Ryan Petrello
519d2eebcb remove unnecessary dotfiles 2019-04-23 17:23:15 -04:00
Michael Abashian
f2ea9003a3 Merge pull request #3493 from dsesami/e2e_backport_3.5.0
E2E backport 3.5.0
2019-04-23 16:51:34 -04:00
Christian Adams
64ae7a6e45 analytics table copies formatted as csv now 2019-04-23 16:26:14 -04:00
softwarefactory-project-zuul[bot]
b47be3c8c7 Merge pull request #3749 from AlanCoding/simple_openstack
Use common method for openstack OS_CLIENT_CONFIG_FILE injection

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-23 20:23:05 +00:00
John Hill
2f7ec6ff13 adding a markerfile for e2e runs 2019-04-23 16:01:45 -04:00
Daniel Sami
75065b6407 e2e stability backport for 3.5 2019-04-23 15:45:27 -04:00
softwarefactory-project-zuul[bot]
2cb6104fe4 Merge pull request #3696 from AlanCoding/django2_warnings
Fix Django 2.0 deprecation warnings

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-23 18:09:58 +00:00
softwarefactory-project-zuul[bot]
f90f8ba9a8 Merge pull request #3755 from ansible/noretry
Adding an option to run all the tests without a retry statement

Reviewed-by: awxbot
             https://github.com/awxbot
2019-04-23 17:48:52 +00:00
John Hill
9733fde560 Adding an option to run all the tests without a retry statement 2019-04-23 13:18:34 -04:00
Ryan Petrello
231c76c9cb Merge pull request #3490 from jakemcdermott/fix-3487
handle insights credential lookups for projects
2019-04-23 12:08:43 -04:00
Jake McDermott
311daf10b8 handle insights credential lookups for projects
Although most scm types correspond to an scm credential lookup, insights uses its own
credential type.
2019-04-23 11:44:03 -04:00
Bill Nottingham
4ca4563a19 Update Azure requirements for Ansible stable-2.8 branch. 2019-04-23 10:50:00 -04:00
Ryan Petrello
96183cf9c4 Merge pull request #3484 from ansible/insights-integration
Insights integration
2019-04-23 10:05:00 -04:00
Jeff Bradberry
0b555e938a Fix typo in the Insights inventory url 2019-04-23 09:39:08 -04:00
AlanCoding
c36dbb3448 Simplify gce inventory plugin injector
This consumes the change made in Ansible core
https://github.com/ansible/ansible/pull/54407
which is in Ansible 2.8, allowing the plugin
injection logic to share the script logic and
to be simplified
2019-04-23 09:33:17 -04:00
AlanCoding
adfce6edf1 Unify openstack inventory injection logic
Remove logic specific to job runs to create a "clouds" file
  for openstack credential type

Move that logic into the collection of managed_by_tower
  injector methods, so it will be used by all job types

Modify inventory openstack injector logic to use this
  data as a base for its logic building the clouds file
2019-04-23 07:59:47 -04:00
AlanCoding
140394fe1f Change credential pattern for openstack inventory plugin
See upstream docs https://github.com/ansible/ansible/pull/54532
Previously it was thought that the entry in the inventory config
file was necessary, but the upstream docs change allows us to
use the same pattern that we used to for the script.
2019-04-22 14:32:00 -04:00
Jeff Bradberry
6ff539e6ee Update the front-end Insights urls 2019-04-22 14:17:47 -04:00
Jeff Bradberry
49ba6c6b3d Appease flake8 2019-04-22 14:17:47 -04:00
Jeff Bradberry
e2861c6c39 Fix the tests to conform to the new Insights results 2019-04-22 14:17:47 -04:00
Jeff Bradberry
6dae4a1d6d Add the host in as a parameter to the _get_insights call 2019-04-22 14:17:47 -04:00
Jeff Bradberry
e66f9241a9 Remove the platform_id from the call to _get_insights
since it is now acquired within _get_insights.
2019-04-22 14:17:47 -04:00
Jeff Bradberry
72da961550 Conform to the new output of the Insights system reports endpoint 2019-04-22 14:17:47 -04:00
Jeff Bradberry
4c86c5065c Fix a typo in the per-system remediations api call 2019-04-22 14:17:47 -04:00
Jeff Bradberry
80a855c57a Do some basic slugification of the remediation playbook name 2019-04-22 14:17:47 -04:00
Jeff Bradberry
11b85250e8 Update more urls 2019-04-22 14:17:47 -04:00
Jeff Bradberry
10cfac2f0e Update the error message when we can't discover the platform ID 2019-04-22 14:17:47 -04:00
Jeff Bradberry
355a83441a Guard against the case where Insights fails to find the system ID 2019-04-22 14:17:47 -04:00
Jeff Bradberry
e25adca233 Replace the old unit tests with new functional tests 2019-04-22 14:17:47 -04:00
Jeff Bradberry
7c743904b0 Removed some no longer needed imports 2019-04-22 14:17:47 -04:00
Jeff Bradberry
a97865de0c Refactor HostInsights
for better reuse of the error handling of the Insights API calls.
2019-04-22 14:17:47 -04:00
Jeff Bradberry
596a5173ce Modify filter_insights_api_response to take in the separate remediations
since it is accumulated via a different API call.
2019-04-22 14:17:47 -04:00
Jeff Bradberry
63209197dd Iterate over the pages of remediations available 2019-04-22 14:17:47 -04:00
Jeff Bradberry
af2484cd97 Update the Insights API urls to use the new url structure 2019-04-22 14:17:47 -04:00
Jeff Bradberry
f01a936202 Update the test for filter_insights_api_response()
and fix the data nesting issue that it uncovered.
2019-04-22 14:17:47 -04:00
Jeff Bradberry
05f670a6d9 Update the filter_insights_api_response() utility function
in order to conform the output of the new Insights Advisor report
endpoint to our expections.
2019-04-22 14:17:47 -04:00
Jeff Bradberry
f4b0910e98 Call out to the Insights API to get the Platform ID for a host
Since the new reports endpoint requires that, rather than the Insights
system ID that we've been using (and storing on the Host model).
2019-04-22 14:17:47 -04:00
Jeff Bradberry
95b17892ee Factor out the response handling boilerplate from HostInsights.get_insights
We'll want to reuse it for the API call to /hosts as well.
2019-04-22 14:17:47 -04:00
Jeff Bradberry
8fdc53cb21 Update the Insights API endpoint for getting the reports for a host 2019-04-22 14:17:47 -04:00
Jeff Bradberry
04521f5c5c Update the Insights action plugin to make calls against the new API 2019-04-22 14:17:47 -04:00
AlanCoding
8c2b3e9b84 Fix Django 2.0 deprecation warnings 2019-04-22 14:17:14 -04:00
softwarefactory-project-zuul[bot]
864fef4f29 Merge pull request #3748 from ryanpetrello/pin-runner-133
pin runner to 1.3.3

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-22 17:28:16 +00:00
Ryan Petrello
787e369bcf pin runner to 1.3.3 2019-04-22 12:57:55 -04:00
softwarefactory-project-zuul[bot]
805a0c7a9a Merge pull request #3730 from jbradberry/manage-org-resource-roles
Adjust the access logic for settings.MANAGE_ORGANIZATION_AUTH = False

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-22 16:44:05 +00:00
softwarefactory-project-zuul[bot]
c80a5e2164 Merge pull request #3747 from ryanpetrello/d-d-d-danger-zone
add a loud warning about perf regressions to tasks.py

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-22 15:25:46 +00:00
Ryan Petrello
eb0463890c add a loud warning about perf regressions to tasks.py 2019-04-22 10:26:58 -04:00
softwarefactory-project-zuul[bot]
f6076052bd Merge pull request #3737 from wenottingham/computing-is-deprecated--tell-your-friends
Deprecate a bunch of inventory computed fields.

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-18 21:11:50 +00:00
Bill Nottingham
0d18d46ccc Sync docker-compose dockerfile with image build dockerfile.
It was installing an IUS repo, rabbitmq-server, and other unneeded things.
2019-04-18 16:50:38 -04:00
softwarefactory-project-zuul[bot]
23ea419aa9 Merge pull request #3724 from ansible/spinnycheck
prevent flake for user e2e

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-18 20:47:00 +00:00
Daniel Sami
b084622c9e prevent flake for user e2e 2019-04-18 16:25:35 -04:00
Bill Nottingham
ef7e1afa34 Add migrations for field names. 2019-04-18 15:39:47 -04:00
softwarefactory-project-zuul[bot]
be86086134 Merge pull request #3739 from elyezer/update-jobs-portal-e2e
Update jobs portal list actions e2e

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-18 19:30:56 +00:00
Elyézer Rezende
a06d2946b6 Update jobs portal list actions e2e 2019-04-18 15:09:29 -04:00
Jeff Bradberry
0ba87c9729 Add more test checks for the alternate code path to the role checks 2019-04-18 14:53:19 -04:00
Jeff Bradberry
41b476544d Improve test coverage of attaching a user to an organization 2019-04-18 13:35:35 -04:00
softwarefactory-project-zuul[bot]
97a6255531 Merge pull request #3736 from ryanpetrello/red-means-loud-on-the-internet
make deprecation warnings at /api/ much more obvious

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-18 17:05:33 +00:00
Ryan Petrello
4159a9b6af make deprecation warnings at /api/ much more obvious 2019-04-18 12:32:49 -04:00
softwarefactory-project-zuul[bot]
798bc4b8de Merge pull request #3734 from wenottingham/sqlitis-is-bad
remove sqlite db, add it to .gitignore

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-18 16:28:40 +00:00
Bill Nottingham
6adf4ca2c1 Deprecate a bunch of inventory computed fields.
Cribbed from https://github.com/ansible/awx/pull/3281/
2019-04-18 12:12:50 -04:00
softwarefactory-project-zuul[bot]
8cc18f501e Merge pull request #3733 from ryanpetrello/dashboard-deprecation
deprecate /api/v2/dashboard in favor of /api/v2/metrics

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-18 16:09:19 +00:00
Bill Nottingham
63049fe99d remove sqlite db, add it to .gitignore 2019-04-18 12:03:04 -04:00
Ryan Petrello
d6ecb486d2 deprecate /api/v2/dashboard in favor of /api/v2/metrics 2019-04-18 11:45:37 -04:00
Vismay Golwala
f4dc4d5849 Cap page_size in pagination urls
Currently, even with a `max_page_size` of n, we can see urls
formed in pagination with `page_size` > n. API still caps the
number of results it returns, but the URL remain invalid there.
This is a bit messy solution to make string replacement in URL
if the query param exceeds `max_page_size`

Signed-off-by: Vismay Golwala <vgolwala@redhat.com>
2019-04-18 11:39:22 -04:00
softwarefactory-project-zuul[bot]
87d55b13bc Merge pull request #3549 from vismay-golwala/verify_venv_on_launch
Validate virtual environment while running a job/inventory update

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-18 12:31:25 +00:00
Vismay Golwala
5d570a017a Validate virtual environment while running a job/inventory update
Currently we only check the custom virtual environment path when
it's created. However, to tackle with the case when the venv might
have been changed/deleted afterward, we need to validate it at
run-time too.

Signed-off-by: Vismay Golwala <vgolwala@redhat.com>
2019-04-18 08:09:49 -04:00
softwarefactory-project-zuul[bot]
6d5897f371 Merge pull request #3728 from chrismeyersfsu/speed_boost
instantiate dispatcher once per job run

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-18 11:14:32 +00:00
softwarefactory-project-zuul[bot]
e14f19468b Merge pull request #3729 from rooftopcellist/rename_analytics_setting
rename analytics setting

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-17 22:55:43 +00:00
Jeff Bradberry
8ad0b3f787 Check the permissions for adding users to orgs/teams in the other direction 2019-04-17 17:45:20 -04:00
Jeff Bradberry
70b0679a0c Adjust the access logic for settings.MANAGE_ORGANIZATION_AUTH = False
so that changing the membership of Organizations and Teams are
disallowed unless you are a superuser, but granting resource
privileges is still permitted.
2019-04-17 15:37:02 -04:00
Christian Adams
add8673d7c rename analytics setting 2019-04-17 15:28:15 -04:00
chris meyers
84c09a19d1 instantiate dispatcher once per job run
* Instantiating the callback dispatch queue on each job event callback
is expensive. Instead, instantiate it only once. Note, we do not need to
instantiate the callback queue in the iso case so we do not.
2019-04-17 14:03:00 -04:00
softwarefactory-project-zuul[bot]
fbc7d1a9f2 Merge pull request #3711 from AlanCoding/conservative_upgrades3
Apply critical dependency upgrades (try 2)

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-17 13:04:24 +00:00
softwarefactory-project-zuul[bot]
daeeb31590 Merge pull request #3715 from ryanpetrello/local-python
properly set ansible_python_interpreter for local task execution

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-17 12:44:39 +00:00
softwarefactory-project-zuul[bot]
66886fb57a Merge pull request #3721 from AlanCoding/org_members_read_teams
Use querset special case to let org members see teams

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-17 12:24:49 +00:00
AlanCoding
1ddb675fa2 Use querset special case to let org members see teams 2019-04-17 07:05:25 -04:00
softwarefactory-project-zuul[bot]
4fd04e095f Merge pull request #2860 from AlanCoding/auditor_association
Show entry for system auditor associations

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-17 10:48:37 +00:00
Ryan Petrello
9d000a76de properly set ansible_python_interpreter for local task execution
this works a limitation in ansible for connection: local and makes it so
that you can run connection: local playbooks with a /usr/bin/python that
is Python3 and an Ansible virtualenv that is Python2

see: https://github.com/ansible/awx/issues/3267
2019-04-16 21:24:23 -04:00
softwarefactory-project-zuul[bot]
7f5227809f Merge pull request #3723 from ryanpetrello/django-works-in-mysterious-ways
fix a bug that breaks the isolated heartbeat

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-16 20:46:14 +00:00
Ryan Petrello
f1d87bf392 fix a bug that breaks the isolated heartbeat 2019-04-16 16:24:40 -04:00
AlanCoding
941009bf6d Apply critical dependency upgrades
fix PyYAML warnings in unit tests
update paramiko source file for license
2019-04-16 16:00:09 -04:00
AlanCoding
9c71204435 show activity stream entry for system auditor association 2019-04-16 15:59:04 -04:00
softwarefactory-project-zuul[bot]
15ef095366 Merge pull request #3720 from shanemcd/memcache-config
Fix memcached configuration in local Docker installs

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-16 19:53:23 +00:00
softwarefactory-project-zuul[bot]
c2daccade7 Merge pull request #3667 from chrismeyersfsu/delete-system-tracking
remove system tracking

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-16 17:24:03 +00:00
Shane McDonald
3bfb54d2fd Fix memcached configuration in local Docker installs
Related: https://github.com/ansible/awx/issues/3719
Signed-off-by: Shane McDonald <me@shanemcd.com>
2019-04-16 12:51:28 -04:00
softwarefactory-project-zuul[bot]
a5d31e56d6 Merge pull request #3709 from AlanCoding/piparoo
Apply same pip and setuptools to AWX and Ansible venv

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-15 21:05:49 +00:00
softwarefactory-project-zuul[bot]
d222bed932 Merge pull request #3712 from jladdjr/iso_node_healthcheck_should_not_reset_capacity
Do not reset capacity of iso nodes when disabled

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-15 20:40:01 +00:00
softwarefactory-project-zuul[bot]
45cc8f1cc9 Merge pull request #3703 from mabashian/upgrade-bootstrap-yay
Bumps Bootstrap to 4.3.1

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-15 20:35:03 +00:00
Vismay Golwala
e0c4fd4b3a Disallow deleting controller or isolated instance groups
Added two new properties to the InstanceGroup model - `is_controller`
and `is_isolated`. Used these properties to hide the trash icon for
instance groups that are either controller or isolated.

Signed-off-by: Vismay Golwala <vgolwala@redhat.com>
2019-04-15 16:08:27 -04:00
Jim Ladd
6ef3b18803 Do not reset capacity of iso nodes when disabled 2019-04-15 12:36:15 -07:00
softwarefactory-project-zuul[bot]
f23b4e7b9a Merge pull request #3708 from rooftopcellist/update_prom_docs
add url to prom docs

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-15 18:23:19 +00:00
softwarefactory-project-zuul[bot]
1d4773545e Merge pull request #3707 from ryanpetrello/report-ansible-runner-crashes
if runner crashes, attempt to record why

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-15 18:16:09 +00:00
AlanCoding
e40c6da552 Apply same pip and setuptools to AWX and Ansible venv 2019-04-15 13:54:25 -04:00
Christian Adams
8ef81065b7 add url to prom docs 2019-04-15 13:39:55 -04:00
mabashian
08fcdf0e25 Bumps Bootstrap to 4.3.1 2019-04-15 13:20:00 -04:00
Ryan Petrello
387682ed8d if runner crashes, attempt to record why
this attempts to surface the underlying runner exception for tracebacks
like this one:

FileNotFoundError: [Errno 2] No such file or directory:
'/tmp/awx_41_93gtgv25/artifacts/41/status'
2019-04-15 13:17:45 -04:00
softwarefactory-project-zuul[bot]
707c3883a8 Merge pull request #3702 from AlanCoding/revert_upgrades
Revert "Merge pull request #3697 from AlanCoding/conservative_upgrades"

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-15 16:54:36 +00:00
softwarefactory-project-zuul[bot]
47b0a96e96 Merge pull request #3704 from rooftopcellist/cleanup_prom_files
add prometheus data to gitignore

Reviewed-by: awxbot
             https://github.com/awxbot
2019-04-15 15:17:07 +00:00
Christian Adams
fb4495b1b5 add prometheus data to gitignore 2019-04-15 10:43:28 -04:00
AlanCoding
5dbc269de1 Revert "Merge pull request #3697 from AlanCoding/conservative_upgrades"
This reverts commit 890de400e2, reversing
changes made to e8de7bc845.
2019-04-15 10:39:56 -04:00
softwarefactory-project-zuul[bot]
5c7939a6ac Merge pull request #3676 from keithjgrant/3604-search-tag-line-wrap
fix tag word wrapping for firefox

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-15 13:56:40 +00:00
softwarefactory-project-zuul[bot]
890de400e2 Merge pull request #3697 from AlanCoding/conservative_upgrades
Apply critical dependency upgrades

Reviewed-by: Alan Rominger <arominge@redhat.com>
             https://github.com/AlanCoding
2019-04-15 13:24:25 +00:00
softwarefactory-project-zuul[bot]
e8de7bc845 Merge pull request #3673 from rooftopcellist/metrics_uuids
add uuids to ping and metrics

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-13 11:45:47 +00:00
Christian Adams
40393e201f add uuids to ping and metrics 2019-04-12 18:37:40 -04:00
softwarefactory-project-zuul[bot]
97e2137d07 Merge pull request #3694 from ryanpetrello/gbye-callback-plugins
remove the old callback plugin import paths

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-12 21:01:34 +00:00
softwarefactory-project-zuul[bot]
29c3927a16 Merge pull request #3695 from wenottingham/words-are-hard
Update wording.

Reviewed-by: Ryan Petrello
             https://github.com/ryanpetrello
2019-04-12 20:35:46 +00:00
Ryan Petrello
17a803f49c remove the old callback plugin import paths and callback-specific tests 2019-04-12 16:11:23 -04:00
softwarefactory-project-zuul[bot]
d0c3882d9d Merge pull request #3698 from ansible/bumpse
Adding selenium stability fix and unpinning se images

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-12 20:10:57 +00:00
Bill Nottingham
9da5cc9c23 Update wording. 2019-04-12 16:09:30 -04:00
softwarefactory-project-zuul[bot]
7913247eaa Merge pull request #3678 from jlmitch5/showRevAndIgForJobResults
subscribe to updates for scm revision and and instance group for job …

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-12 19:53:01 +00:00
AlanCoding
569cd3b34e update paramiko source file 2019-04-12 15:52:35 -04:00
AlanCoding
2c67bec3cb fix PyYAML warnings in unit tests 2019-04-12 15:48:08 -04:00
John Hill
77e587babe Adding selenium stability fix and unpinning se images 2019-04-12 15:43:37 -04:00
AlanCoding
235011f5e9 Apply critical dependency upgrades 2019-04-12 15:26:48 -04:00
chris meyers
0c6a522813 remove system tracking
* Leave the artisanal handcrafted migration from vendored fact scan to
external fact gathering.
2019-04-12 14:30:29 -04:00
softwarefactory-project-zuul[bot]
dc833bbaa7 Merge pull request #3683 from mabashian/3565-workflow-template-popover
Leverages awPopOverWatch to generate popover dynamically on workflow node templates help

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-12 16:54:54 +00:00
mabashian
9b0329021c Leverages awPopOverWatch to generate popover dynamically on workflow node templates help 2019-04-12 12:29:50 -04:00
softwarefactory-project-zuul[bot]
3611f3491b Merge pull request #3581 from beeankha/basic_license_feature
Update Basic License Feature Access

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-12 15:57:18 +00:00
softwarefactory-project-zuul[bot]
213a0c3cef Merge pull request #3692 from ansible/spinnycheck
Fixing flake and waits for spinny that don't need to be there.

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-12 15:45:44 +00:00
beeankha
0cc640403a Resolve initial login error 2019-04-12 11:28:27 -04:00
Ryan Petrello
85898fd708 remove more unnecessary feature flagging for AUTHENTICATION_BACKENDS 2019-04-12 11:28:27 -04:00
beeankha
14d4d624e4 Restore button for adding notifications 2019-04-12 11:28:27 -04:00
beeankha
de34a64115 Basic License feature gating changes 2019-04-12 11:28:27 -04:00
Daniel Sami
d67a336e2f drop spinny
flake fixes

adjust
2019-04-12 10:57:35 -04:00
softwarefactory-project-zuul[bot]
58966d7368 Merge pull request #3625 from ryanpetrello/iso-forks
WIP: specify --forks on isolated health check calls

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-11 21:41:37 +00:00
softwarefactory-project-zuul[bot]
fc5322b2a4 Merge pull request #3675 from mabashian/as-mgmt-jobs
Check to make sure the activity stream button should be shown after refresh

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-11 17:54:24 +00:00
softwarefactory-project-zuul[bot]
817b350de9 Merge pull request #3685 from ryanpetrello/runner-132
pin ansible runner to 1.3.2

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-11 17:27:51 +00:00
Ryan Petrello
0c4ed78bee pin ansible runner to 1.3.2 2019-04-11 12:57:07 -04:00
softwarefactory-project-zuul[bot]
0252af0d30 Merge pull request #3674 from ansible/splits
added ws and pending to user e2e

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-11 16:04:34 +00:00
Daniel Sami
f56f6c11f7 added ws and pending to user e2e
lint
2019-04-11 09:41:51 -04:00
John Mitchell
ef84bfc1c2 subscribe to updates for scm revision and and instance group for job results 2019-04-10 17:00:35 -04:00
softwarefactory-project-zuul[bot]
8383568e3c Merge pull request #3672 from rooftopcellist/metrics_content_types
remove invalid json renderer from metrics endpoint

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-10 20:36:42 +00:00
Christian Adams
fd5d9c45d7 remove invalid json renderer from metrics endpoint 2019-04-10 16:14:54 -04:00
softwarefactory-project-zuul[bot]
23d0f6022c Merge pull request #3671 from athenahealth/fix-awx-logos-inventory-comments
Fix comment on awx branding inventory var

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-10 20:05:16 +00:00
mabashian
9e98058290 Check to make sure the activity stream button should be shown after refresh 2019-04-10 15:56:19 -04:00
Keith Grant
e073b5e017 fix tag word wrapping for firefox 2019-04-10 15:51:56 -04:00
softwarefactory-project-zuul[bot]
21bec83a4e Merge pull request #3669 from mabashian/workflow-node-save
Ensures extra vars are converted to yaml before being passed along to prompt steps

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-10 19:50:39 +00:00
softwarefactory-project-zuul[bot]
cd966f2669 Merge pull request #3670 from mabashian/root-all-groups-responsive
Make root/all groups action bar a bit more responsive

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-10 19:50:33 +00:00
softwarefactory-project-zuul[bot]
1830da4268 Merge pull request #3656 from elyezer/applications-e2e
Add applications to e2e

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-10 18:57:22 +00:00
Jeff Byrnes
98ec5c8250 Fix comment on awx branding inventory var
Fix a conflict with the “AWX Branding” in INSTALL.md, which
has the correct instructions.
2019-04-10 14:49:37 -04:00
mabashian
0caf1e8a3d Make root/all groups action bar a bit more responsive 2019-04-10 12:06:47 -04:00
softwarefactory-project-zuul[bot]
ae7b173e17 Merge pull request #3661 from AlanCoding/stdout_cleanup
Remove stdout wrapper classes that moved to runner

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-10 15:05:32 +00:00
softwarefactory-project-zuul[bot]
c326b186a6 Merge pull request #3629 from AlanCoding/nt_read_enable
Fix RBAC bugs with notification attachment

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-10 14:56:37 +00:00
mabashian
4f2443e7bb Ensures extra vars are converted to yaml before being passed along to prompt steps 2019-04-10 10:51:27 -04:00
Elyézer Rezende
213a70b98a Add applications to e2e 2019-04-10 10:50:49 -04:00
mabashian
c91cd606ed Corresponding UI changes for notifications tab and toggle permissions 2019-04-10 10:31:31 -04:00
AlanCoding
4eab362318 fix RBAC bugs with notification attachment
Allow notification_admin_role users to attach
  NTs from that organization
Require either read_role or auditor_role to the
  object which the NT is being attached to
2019-04-10 10:29:54 -04:00
softwarefactory-project-zuul[bot]
df81eb7533 Merge pull request #3659 from rooftopcellist/prom_test
add prom permissions test

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-10 14:24:42 +00:00
softwarefactory-project-zuul[bot]
1bb5d560bd Merge pull request #3663 from kialam/fix-3414-job-output-live-updates-disabled
Fix job output name in websocket service.

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-10 13:52:08 +00:00
softwarefactory-project-zuul[bot]
652281b6fa Merge pull request #3645 from AlexSCorey/2281-HostEventToolTip
Add tool tip to Host Event lines.

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-09 21:17:08 +00:00
Kia Lam
b978a28f8b Fix job output name in websocket service. 2019-04-09 17:04:30 -04:00
softwarefactory-project-zuul[bot]
4664d9556d Merge pull request #3620 from AlanCoding/gce_instance_id
Add instance_id for gce imported hosts

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-09 20:49:06 +00:00
softwarefactory-project-zuul[bot]
7f935084df Merge pull request #3653 from AlanCoding/instance_token_management2
Fix 403 bug using write PAT and view.always_allow_superuser=True

Reviewed-by: Alan Rominger <arominge@redhat.com>
             https://github.com/AlanCoding
2019-04-09 20:37:05 +00:00
AlanCoding
988438f119 Remove stdout wrapper classes that moved to runner 2019-04-09 16:28:30 -04:00
softwarefactory-project-zuul[bot]
20a023e243 Merge pull request #3658 from keithjgrant/2526-fix-workflow-results-layout
Fix workflow results layout when extra vars has long line

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-09 19:40:26 +00:00
Christian Adams
f49532090b add prom permissions test 2019-04-09 15:12:04 -04:00
Alex Corey
75a1c69ea2 address concerns in PR 2019-04-09 14:55:34 -04:00
softwarefactory-project-zuul[bot]
7f6659f767 Merge pull request #3657 from ryanpetrello/issue-template-remove-component
remove COMPONENT NAME from our issue templates

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-09 17:50:30 +00:00
Keith Grant
8f05482f8f fix workflow results layout when extra vars has long line 2019-04-09 13:48:25 -04:00
Ryan Petrello
01a79dd23e remove COMPONENT NAME from our issue templates
this is a remnant from our closed sourced days; upstream bug reporters
only fill this out correctly like 50% of the time, and it's often
difficult to know without knowing how AWX actually works
2019-04-09 12:49:11 -04:00
softwarefactory-project-zuul[bot]
3d308cc2a3 Merge pull request #3636 from ansible/prometheus
Prometheus Metrics

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-09 15:51:27 +00:00
Alex Corey
4f55ffe4a0 Add tool tip to Host Event lines. 2019-04-09 10:58:23 -04:00
softwarefactory-project-zuul[bot]
6ce972a2a4 Merge pull request #3652 from keithjgrant/3639-codemirror-popout-bug
Don't obscure modal buttons when code mirror modal throws error

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-09 14:20:21 +00:00
softwarefactory-project-zuul[bot]
57b3565f42 Merge pull request #3631 from mabashian/reload-on-pendo-change
Refresh the page after pendo flag changes to make change effective immediately

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-09 14:13:00 +00:00
Wayne Witzel III
1abb0b2c35 restrict metrics to superuser and system auditor 2019-04-09 10:07:38 -04:00
AlanCoding
809fcac738 fix 403 bug using write PAT and view.always_allow_superuser=True 2019-04-09 09:54:52 -04:00
Keith Grant
fa665e81e4 don't obscure modal buttons when code mirror modal throws error 2019-04-09 09:30:58 -04:00
AlanCoding
c4b41a174b add instance_id for gce imported hosts 2019-04-09 08:20:26 -04:00
softwarefactory-project-zuul[bot]
ce65fdd26c Merge pull request #3647 from ryanpetrello/cred-namespace-api
expose CredentialType.namespace in the API

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-09 11:50:01 +00:00
Ryan Petrello
aa0179690d expose CredentialType.namespace in the API
see: https://github.com/ansible/awx/issues/3644
2019-04-08 19:02:10 -04:00
softwarefactory-project-zuul[bot]
67aeecdee5 Merge pull request #3637 from ryanpetrello/install-uuid
add a unique UUID for identifying an AWX installation

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-08 20:17:19 +00:00
softwarefactory-project-zuul[bot]
d14b71ccee Merge pull request #3643 from AlanCoding/cloudforms_py3
Backport fix to run cloudforms script in py3

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-08 18:06:49 +00:00
AlanCoding
47a3ee7d76 backport fix to run cloudforms script in py3 2019-04-08 13:40:44 -04:00
softwarefactory-project-zuul[bot]
a2bfb0f65c Merge pull request #3633 from AlanCoding/slice_of_one
Avoid slicing if the inventory only has 1 host

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-08 16:57:51 +00:00
Ryan Petrello
fc9da002d2 add an example config file and make target for starting a prometheus 2019-04-08 12:01:46 -04:00
Christian Adams
e1c6057b4c add insights setting to metrics 2019-04-08 11:49:41 -04:00
Wayne Witzel III
520cbd2015 update prometheus run example 2019-04-08 11:47:10 -04:00
Christian Adams
e2039b7d3f add insights setting to metrics 2019-04-08 11:34:38 -04:00
softwarefactory-project-zuul[bot]
9125313ed3 Merge pull request #3640 from konkolorado/devel
Removes failing symlink to /usr/bin/python3 during the image build

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-08 14:30:49 +00:00
Wayne Witzel III
2c8900568b add prometheus-client license details 2019-04-08 10:21:05 -04:00
Wayne Witzel III
5c1d2a6f0b flake8 cleanup 2019-04-08 09:35:46 -04:00
Wayne Witzel III
bb5c7a98f3 test prometheus metrics output 2019-04-08 09:28:57 -04:00
Christian Adams
3fb3079264 fix job status metric 2019-04-08 02:47:56 -04:00
Wayne Witzel III
c3812de3d6 initial prometheus commit
Co-authored-by: Wayne Witzel III <wayne@riotousliving.com>
Co-authored-by: Christian Adams <rooftopcellist@gmail.com>
2019-04-08 02:47:56 -04:00
Uriel Mandujano
4821a94944 Removes failing symlink to /usr/bin/python3 during the image build
Installing the latest python36-setuptools automatically creates the symlink from python3 -> python3.6 and from python36 -> /usr/bin/python3.6. Building the images fails when the symlink is created explicitly in the AWX installer.

Signed-off-by: Uriel Mandujano <uriel.mandujano14@gmail.com>
2019-04-06 11:13:41 -05:00
Uriel Mandujano
97e030dd1f Revert "Removes failing symlink to /usr/bin/python3 because that file already exists"
This reverts commit 13fadd3838.
2019-04-06 11:10:35 -05:00
Ryan Petrello
dfd4cb55e5 add a unique UUID for identifying an AWX installation 2019-04-05 20:10:55 -04:00
Uriel Mandujano
13fadd3838 Removes failing symlink to /usr/bin/python3 because that file already exists 2019-04-05 16:53:20 -05:00
Jeff Byrnes
7b636a7566 Set up HTTPS w/ proper port & HTTP redirect
HTTPS is, by default, expected to be on port 443.

Also, with HSTS set, we need to be sure that users attempting to arrive
via HTTP are properly redirected to HTTPS.

This does so by:

* Setting up a 301 redirect for any URL to its HTTPS version
* Adjusting the internal port for HTTPS traffic to 8053
* Setting docker-compose to share port 443 → 8053
    - This is configurable via an inventory variable
2019-04-05 16:13:23 -04:00
Jeff Byrnes
28e3c63562 Add optional SSL cert to docker-compose install
In #3322, this mount was added, but only to the standalone
Docker install setup:

github.com/ansible/awx/pull/3322/files#diff-596e32ab54a52bfed763f8a639499fe0

This ensures that the SSL cert is loaded when using docker-compose,
which is the only Docker-based method available as of v4.0.0
2019-04-05 16:13:23 -04:00
AlanCoding
cca9de9a3e set default slice ct for special cases 2019-04-05 16:12:40 -04:00
AlanCoding
880341ac05 avoid slicing if the inventory only has 1 host 2019-04-05 15:28:29 -04:00
softwarefactory-project-zuul[bot]
e9f2fddc7f Merge pull request #3635 from ryanpetrello/i-am-so-smrt
fix a typo

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-05 18:57:16 +00:00
Ryan Petrello
a9a479a51b fix a typo 2019-04-05 14:28:39 -04:00
softwarefactory-project-zuul[bot]
ddcb7d4881 Merge pull request #3634 from ryanpetrello/dashboard-confessionals
fix a small bug related to failed inventory counts in the dashboard API

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-05 18:24:39 +00:00
Ryan Petrello
cd90ad2497 fix a small bug related to failed inventory counts in the dashboard API 2019-04-05 14:02:19 -04:00
softwarefactory-project-zuul[bot]
e3dfc6c796 Merge pull request #3596 from jbradberry/capture-isolated-command
Updated IsolatedManager to take a callback that captures the remote command

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-05 17:15:11 +00:00
softwarefactory-project-zuul[bot]
1a151ad63a Merge pull request #3587 from chrismeyersfsu/fix-pickup_insights_id_from_ansible_facts
pickup insights system_id from ansible facts

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-05 16:07:21 +00:00
softwarefactory-project-zuul[bot]
4095a6c9d4 Merge pull request #3632 from ryanpetrello/duplicate-playbook-stdout
don't write playbook stdout to sys.stdout (it's duplicated in log files)

Reviewed-by: awxbot
             https://github.com/awxbot
2019-04-05 15:50:17 +00:00
softwarefactory-project-zuul[bot]
161fd9cde0 Merge pull request #3460 from vismay-golwala/project_admin_issue
Project admin manual SCM Type creation bug fix

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-05 15:43:49 +00:00
softwarefactory-project-zuul[bot]
dc77ddbc5b Merge pull request #3617 from mabashian/jt-enter-toggle
Prevent on/off toggles from getting inadvertently trigged by enter key presses

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-05 15:31:02 +00:00
Ryan Petrello
81fe923577 don't write playbook stdout to sys.stdout (it's duplicated in log files)
this instructs runner to _not_ write to stdout when we invoke
runner.interface.run(); AWX consumes/ingests this strictly as events
2019-04-05 11:20:34 -04:00
softwarefactory-project-zuul[bot]
3521c56baf Merge pull request #3608 from mabashian/host-filter-quotes
Prevents replacing encoded quotes while searching against host filter

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-05 15:14:29 +00:00
softwarefactory-project-zuul[bot]
dd609b8a7c Merge pull request #3622 from wenottingham/roleing-roleing-roleing
Adjust descriptions of RBAC roles for clarity.

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-05 15:09:42 +00:00
softwarefactory-project-zuul[bot]
ef1a5c09b6 Merge pull request #3510 from jbradberry/errors-on-change-password
Use Django's own logic to invalidate sessions of users when changing passwords

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-05 14:54:20 +00:00
mabashian
fa3a41f25b Refresh the page after pendo flag changes to make change effective immediately 2019-04-05 10:48:53 -04:00
softwarefactory-project-zuul[bot]
50c7807483 Merge pull request #3628 from marshmalien/fix-org-user-linkout
Fix Organization User badge linkout

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-05 14:37:53 +00:00
chris meyers
531d97d3b3 pickup insights system_id from ansible facts
* Continue to pick up facts from scan_insights.py
* This PR adds the ability to pickup facts from
/etc/ansible/facts.d/insights.facts
* Log what transport the insights system_id was found via
2019-04-05 10:35:56 -04:00
softwarefactory-project-zuul[bot]
d4c69429db Merge pull request #3624 from ryanpetrello/iso-cleanup-path-changes
update periodic isolated cleanup to match the new paths post-runner

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-05 14:25:18 +00:00
Ryan Petrello
79d580d5b9 update periodic isolated cleanup to match the new paths post-runner 2019-04-05 09:43:27 -04:00
Marliana Lara
8302992a35 Fix Organization User badge linkout 2019-04-05 08:44:12 -04:00
softwarefactory-project-zuul[bot]
4f8b197b55 Merge pull request #3615 from ryanpetrello/platform-metrics
add some additional analytics metrics

Reviewed-by: Ryan Petrello
             https://github.com/ryanpetrello
2019-04-05 00:50:04 +00:00
Ryan Petrello
5a4a812c73 specify --forks on isolated health check calls
this requires ansible-runner 1.3.2
2019-04-04 20:12:14 -04:00
Ryan Petrello
5c5173956d add some additional analytics metrics 2019-04-04 20:10:54 -04:00
softwarefactory-project-zuul[bot]
2423d9df44 Merge pull request #3626 from ryanpetrello/no_ln_py3
fix busted CI

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-04 23:15:42 +00:00
Ryan Petrello
ce2fc1a9dd fix more py3 centos7 issues 2019-04-04 18:35:59 -04:00
Matthew Jones
fd3a423b07 The python 3.6 package seems to already make this symlink 2019-04-04 18:35:56 -04:00
Jeff Bradberry
c6643946c5 Capture the redacted credential env vars separately
and then make use of them specifically to make safe the env vars
coming back from an isolated node.  This will allow us to capture the
safed versions of custom credential values, but without potentially
clobbering normal env var values that vary between the controller and
the node.
2019-04-04 15:22:27 -04:00
softwarefactory-project-zuul[bot]
6b422d3bb7 Merge pull request #3619 from ansible/headless
headless option added

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-04 18:46:32 +00:00
Jeff Bradberry
3f6d3506c6 Change the artifact file convention for isolated nodes to 'command'
since that's what landed in the ansible-runner PR.
2019-04-04 14:25:50 -04:00
Daniel Sami
ac11b3782b headless profile 2019-04-04 14:17:53 -04:00
Bill Nottingham
cfdecd7297 Adjust descriptions of RBAC roles for clarity. 2019-04-04 14:08:30 -04:00
softwarefactory-project-zuul[bot]
5debfdf5e8 Merge pull request #3545 from chrismeyersfsu/fix-tower_header_on_insights_playbook_fetches
include tower header to insights on plan fetch

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-04 15:51:06 +00:00
softwarefactory-project-zuul[bot]
3f9fd3a3a8 Merge pull request #3618 from rooftopcellist/dashboard_sync_jobs
don't show implicit update jobs in dashboard

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-04 15:43:39 +00:00
Christian Adams
2493374d44 don't show implicit update jobs in dashboard 2019-04-04 11:04:27 -04:00
mabashian
7722f0ca08 Prevent on/off toggles from getting inadvertently trigged by enter key presses 2019-04-04 10:16:39 -04:00
softwarefactory-project-zuul[bot]
e3f10ebd65 Merge pull request #3501 from beeankha/timeout_upgrade
Upgrade Old Email Timeout Notifications

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-04 13:23:43 +00:00
softwarefactory-project-zuul[bot]
713b2c1bf2 Merge pull request #3603 from athenahealth/fix-some-docker-standalone-bits
Fix some docker standalone bits

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-04 13:11:44 +00:00
softwarefactory-project-zuul[bot]
00a9ae0e72 Merge pull request #3614 from AlexSCorey/2220-AdminBtnRBAC
Removes Add Organization Admin Button

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-04 13:03:43 +00:00
Alex Corey
6bcb471584 Removes Add Organization Admin Button 2019-04-04 08:13:40 -04:00
beeankha
f5ea595763 Enable timeout feature to remain functional after upgrade
Have the UI spinner fill with a default upon notification type change or
upgrade.
2019-04-03 19:06:11 -04:00
softwarefactory-project-zuul[bot]
1f1cb2bdac Merge pull request #3598 from rooftopcellist/index_event_expiry
add index for jobevent created

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-03 20:56:57 +00:00
mabashian
d9d3c5d15f Adds comment explaining logic following the split on empty string in splitFilterIntoTerms 2019-04-03 16:56:36 -04:00
mabashian
47d6e5c028 Prevents replacing encoded quotes while searching against host filter 2019-04-03 16:50:45 -04:00
Jeff Bradberry
467700e4bb Bring the check_callback back into the loop
but try to process it only once.
2019-04-03 16:04:07 -04:00
softwarefactory-project-zuul[bot]
cfa2eabb57 Merge pull request #3576 from marshmalien/feat-toolbar-sort-template-list
Add sort toolbar to template lists

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-03 19:32:38 +00:00
Jeff Bradberry
b4e508f72a Bring the check_callback call out of the loop
We shouldn't need to call it multiple times.
2019-04-03 15:12:29 -04:00
softwarefactory-project-zuul[bot]
2bda81661e Merge pull request #3592 from keithjgrant/code-mirror-fixes
Code Mirror: prevent console errors when CM parses invalid syntax

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-03 18:57:54 +00:00
Jeff Bradberry
b0f6d2214c Fix a typo: there is no method called check_callback on BaseTask 2019-04-03 14:57:02 -04:00
softwarefactory-project-zuul[bot]
c9bac0b51c Merge pull request #3594 from keithjgrant/407-job-launch-keyboard-nav
Job launch keyboard navigation improvements

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-03 18:51:24 +00:00
softwarefactory-project-zuul[bot]
487473f0d1 Merge pull request #3586 from AlanCoding/update_verbosity
Set ansible-inventory verbosity for Ansible 2.8

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-03 18:51:09 +00:00
Jeff Bradberry
4364e00117 Do the env vars redaction for isolated nodes on this side 2019-04-03 14:34:09 -04:00
Christian Adams
e033eb5aef add index for jobevent created 2019-04-03 14:27:22 -04:00
softwarefactory-project-zuul[bot]
bc169fe1cc Merge pull request #3572 from ansible/pagination-stuff
add pagination to user e2e

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-03 17:34:33 +00:00
softwarefactory-project-zuul[bot]
c15502e581 Merge pull request #3601 from ryanpetrello/isolated-newline-strip
fix a bug in isolated OpenSSH key syncing

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-03 16:18:33 +00:00
Daniel Sami
45d5999bc2 add pagination to user e2e
lint
2019-04-03 12:11:28 -04:00
Ryan Petrello
78cd1abbb2 fix a bug in isolated OpenSSH key syncing
OpenSSH keys _must_ end with a \n to be accepted by ssh-add; enforce
a newline if there isn't one
2019-04-03 11:53:45 -04:00
Jeff Byrnes
e0861fee3a Update INSTALL docs re: docker-compose prereqs
As it turns out, the docker-compose Python module is
required, and docker-py doesn’t cut it.

Even more confusing, docker-compose Python module installs
the docker Python module, which conflicts with docker-py. To
avoid this, there are additional docs to call this out.
2019-04-03 11:52:10 -04:00
Jeff Byrnes
cb806b1699 Set docker_compose_dir like other inventory vars
When docker-compose become the sole method for using
Docker directly, some of this was shifted around in ways that
are inconsistent with other elements.

This adjusts it so that:

* The inventory variable default is set like the others, and
is less confusing
* We no longer mention the Standalone Docker in inventory
* We format our INSTALL docs w/r/t this var
2019-04-03 11:52:10 -04:00
softwarefactory-project-zuul[bot]
15b80cdb1a Merge pull request #3574 from AlexSCorey/265-JTLaunchResponsivness
Wraps launch template buttons

Reviewed-by: Alex Corey <Alex.swansboro@gmail.com>
             https://github.com/AlexSCorey
2019-04-03 15:41:43 +00:00
softwarefactory-project-zuul[bot]
bf0f1f1496 Merge pull request #3546 from jlmitch5/fixLaunchButtonPositioning
move wf and jt form launch buttons to the left instead of the right

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-03 14:49:13 +00:00
Keith Grant
dbe4417ac3 fix lint error 2019-04-03 08:59:14 -04:00
Alex Corey
c76e97cecf Wraps launch template buttons 2019-04-03 08:30:51 -04:00
softwarefactory-project-zuul[bot]
ce4e34eb28 Merge pull request #3597 from wenottingham/how-many-crypto-libs-does-one-cloud-stack-need
Don't use the `rsa` python lib.

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-03 00:22:27 +00:00
Jeff Bradberry
32286a9d49 Change the artifact to also capture the actual envvars data 2019-04-02 17:10:26 -04:00
Bill Nottingham
ed0bb127e8 Don't use the rsa python lib. 2019-04-02 15:48:30 -04:00
Jeff Bradberry
cac48e7cfb Updated IsolatedManager to take a callback that captures the remote command 2019-04-02 15:40:56 -04:00
softwarefactory-project-zuul[bot]
96bd0f9d9e Merge pull request #3573 from AlanCoding/custom_empty_default
Change default to allow empty groups from custom sources

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-02 17:54:54 +00:00
Keith Grant
bc7472f3f1 remove unused parameter 2019-04-02 13:54:26 -04:00
Keith Grant
5b2f00b978 make Enter key navigate to next tab in launch modal 2019-04-02 13:26:31 -04:00
softwarefactory-project-zuul[bot]
3f73176ef2 Merge pull request #3098 from ansible/credential_plugins
Credential Plugins

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-02 16:19:00 +00:00
Jake McDermott
d987c6e3f7 mention usage with custom credential types in credential plugins doc 2019-04-02 11:24:59 -04:00
Jake McDermott
5d11400f6c replace unlinked fields before updating input sources
With credential type changes for linked credentials disabled, we can
update the unlinked credential fields first. If an error occurs when
updating the unlinked fields, no input sources will be changed.
2019-04-02 11:24:55 -04:00
Jake McDermott
6f43f223b2 update e2e credential checks 2019-04-02 11:24:52 -04:00
Jake McDermott
6b87241099 render non-editable linked credentials
We still want to show linked credentials even when a field is
non-editable.
2019-04-02 11:24:48 -04:00
Jake McDermott
0f6615d9cd make linked fields look disabled when disabled 2019-04-02 11:24:44 -04:00
Jake McDermott
9b55fa61de fix required field entry for object_query 2019-04-02 11:24:41 -04:00
Jake McDermott
165405ad3f include externally sourced input fields for injection 2019-04-02 11:24:36 -04:00
Jake McDermott
7ca92e4c1e prevent input source changes without use role on source cred
To update an input source, the user must have admin access
to the target credential and at least use role on the source
credential.
2019-04-02 11:24:32 -04:00
Jake McDermott
8c107a5fa8 prevent overflow in plugin test notifications 2019-04-02 11:24:14 -04:00
Jake McDermott
8b35ac89fc add timeouts to plugin backends 2019-04-02 11:24:11 -04:00
Jake McDermott
b90f9ac401 remove linked fields from validation loop
All linked fields are initially valid. All unlinked fields
are initially valid.
2019-04-02 11:24:07 -04:00
Jake McDermott
7a093de9fd add workaround for null cloud credential kind summary fields
job details summary field credentials have `null` for the
`kind` field so we can't map it to icon names. To work around
this, we now check the value of the `cloud` field when generating
the icon to use.
2019-04-02 11:24:04 -04:00
Jake McDermott
259fcdc5df don't initialize linked textarea fields as invalid 2019-04-02 11:24:00 -04:00
Jake McDermott
aa4575c92e update validation when input source is set 2019-04-02 11:23:57 -04:00
Jake McDermott
f0f4f487fb disable input source lookups when they can't be changed
If a user doesn't have permission to change an input field, we disable
the input field.
2019-04-02 11:23:53 -04:00
Jake McDermott
c4a79a778f add delete to input source user_capabilities 2019-04-02 11:23:50 -04:00
Jake McDermott
f611d4275f add use to credential user_capabilities 2019-04-02 11:23:46 -04:00
Jake McDermott
8c3dfe8655 remove input value arg from plugin example
The plugin backends no longer expect an input value as an argument
so we're removing it from the documented examples.
2019-04-02 11:23:42 -04:00
Jake McDermott
f9e081046c add object query fields 2019-04-02 11:23:39 -04:00
Jake McDermott
da9abc087e don't ignore choices param in input config
We always want to use choices if they're available in the input config
from the api. An input can sometimes have no type. Usually we'd throw an
error but we can still load a component if it defines a set of choices
to use instead of a type.
2019-04-02 11:23:35 -04:00
Jake McDermott
956f588fd8 generalize secret textarea hint 2019-04-02 11:23:28 -04:00
Jake McDermott
cb238c1f24 read cert from memory
Certificate data lives in the db. We're using a fifo here so we can avoid writing
it to disk when making the http request.
2019-04-02 11:23:24 -04:00
Jake McDermott
3dee9f0512 add plugin for cyberark aim 2019-04-02 11:23:20 -04:00
Jake McDermott
261a635005 remove non-matches from input source arrays 2019-04-02 11:23:17 -04:00
Jake McDermott
4f13592430 hide drag-and-drop textarea hints when showing input value as tag
You can't drag and drop or enter text when a text area is displaying its value
as a tag, so we hide the hint to avoid incorrectly suggesting that you can.
2019-04-02 11:23:13 -04:00
Jake McDermott
73c9d1b7a9 remove bootstrap fade from dialog component
This was breaking the fixed positioning of help popovers for form inputs
nested within the dialog so I'm replacing it.
2019-04-02 11:23:10 -04:00
Jake McDermott
8180a2060a rename at-easy-modal to at-dialog 2019-04-02 11:23:06 -04:00
Jake McDermott
dfaf19cdf3 use default action button class when fill and color props aren't given 2019-04-02 11:23:02 -04:00
Jake McDermott
ea9ed31f9d refactor metadata conversion function to use reduce 2019-04-02 11:22:59 -04:00
Jake McDermott
05226333ff move tag max height declaration to input tag wrapper
We don't want to apply max height to all tags, just the ones we embed
within text/textarea input fields.
2019-04-02 11:22:55 -04:00
Jake McDermott
5b79843390 use a shared variable for layout declarations 2019-04-02 11:22:51 -04:00
Jake McDermott
1eda939ce2 add tips for secret controls 2019-04-02 11:22:48 -04:00
Jake McDermott
43456d13c4 don't replace input source unless changed 2019-04-02 11:22:44 -04:00
Jake McDermott
215c3c87e5 enable input source linking for secret textarea fields 2019-04-02 11:22:41 -04:00
Jake McDermott
957804e22a enable input source linking for textarea fields 2019-04-02 11:22:37 -04:00
Jake McDermott
e14f17687c disable prompt-on-launch when input source is set 2019-04-02 11:22:34 -04:00
Jake McDermott
e0d61cfb8e enable input source linking for password fields 2019-04-02 11:22:30 -04:00
Jake McDermott
1344706095 add wrapper for text input tags 2019-04-02 11:22:27 -04:00
Jake McDermott
5c855b5bd1 add selected credential tray to input source lookup 2019-04-02 11:22:23 -04:00
Jake McDermott
6d0f2948aa don't show lookup until data is fetched 2019-04-02 11:22:20 -04:00
Jake McDermott
736bd2ed67 add validation for required values and metadata fields 2019-04-02 11:22:16 -04:00
Jake McDermott
47f31b41fb combine add and edit controllers 2019-04-02 11:22:12 -04:00
Jake McDermott
61eeb630f8 move org edit permission check to route resolve 2019-04-02 11:22:08 -04:00
Jake McDermott
7f55a1da0d move input value initialization to models 2019-04-02 11:22:05 -04:00
Jake McDermott
ceef7f57af add input source creation ui 2019-04-02 11:21:59 -04:00
Jake McDermott
393ad6b2f4 add cyberark conjur to tested credential types 2019-04-02 11:21:51 -04:00
Ryan Petrello
0768c6ac1d store the public key for HashiVault signing in the plugin metadata 2019-04-02 11:21:47 -04:00
Ryan Petrello
2824616ba6 add support for CyberArk Conjur (API v5) 2019-04-02 11:21:44 -04:00
Jake McDermott
c436dcf875 add input source prompting and plugin testing 2019-04-02 11:21:40 -04:00
Jake McDermott
0de8a89293 support input source metadata in plugin test apis 2019-04-02 11:21:36 -04:00
Ryan Petrello
81a509424a prefetch related source credentials in tasks.py 2019-04-02 11:21:33 -04:00
Ryan Petrello
42f4956a7f enforce required credential fields at job start time rather than on save
this is necessary for credential plugins support so that you can (in two
requests):

1.  Save a Credential with _no_ input values defined
2.  Create/associate one (or more) CredentialInputSource records to the
    new Credential
2019-04-02 11:21:29 -04:00
Jake McDermott
e2d474ddd2 document restriction of external-external credential source linking 2019-04-02 11:21:25 -04:00
Ryan Petrello
011d7eb892 clean up access to various CredentialInputSource fields (#3336) 2019-04-02 11:21:21 -04:00
Ryan Petrello
018ff91620 add related and summary fields to the CredentialInputSource endpoint 2019-04-02 11:21:17 -04:00
Ryan Petrello
368d933799 remove association behavior from /api/v2/credentials/input_sources/ 2019-04-02 11:21:14 -04:00
Jake McDermott
e9532dea8e cache dynamic input fields
Query dynamic input fields once on attribute access and then cache it for future use.
2019-04-02 11:21:10 -04:00
Ryan Petrello
b911f8bf77 allow creation at /api/v2/credential_input_sources 2019-04-02 11:21:06 -04:00
Ryan Petrello
e727909a61 rename the CredentialInputSource related_names so they're plural 2019-04-02 11:21:02 -04:00
Ryan Petrello
13366c1e75 Encrypt machine.ssh_public_key_data (in case users paste in signed data) 2019-04-02 11:20:58 -04:00
Ryan Petrello
ca6d124417 add API examples for supported credential plugins 2019-04-02 11:20:55 -04:00
Ryan Petrello
35cca68f04 add RBAC definitions for CredentialInputSource 2019-04-02 11:20:51 -04:00
Ryan Petrello
dcf17683e2 mark cred plugin strings for translation 2019-04-02 11:20:48 -04:00
Ryan Petrello
b851e2be4a don't add hvac as a dependency for hashicorp vault integration
hvac is just based on requests anyways, and it doesn't support half of
what we need (like the SSH secrets engine API)
2019-04-02 11:20:44 -04:00
Ryan Petrello
69368d874e move path parameterization to the CredentialInputSource model 2019-04-02 11:20:36 -04:00
Jake McDermott
0ee223f799 add api for testing credential plugins 2019-04-02 11:20:33 -04:00
Ryan Petrello
7a43f00a5d add support for HashiCorp signed SSH certificates 2019-04-02 11:20:29 -04:00
Ryan Petrello
4ed5bca5e3 add credential plugin support for Azure Key Vault 2019-04-02 11:20:26 -04:00
Ryan Petrello
0a87469225 give credential plugins an explicit namespace 2019-04-02 11:20:22 -04:00
Ryan Petrello
63997838cd support HashiCorp Vault versioned secrets (API v2) 2019-04-02 11:20:19 -04:00
Ryan Petrello
89b731a0cb Improve the HashiCorp Vault KV name and field labels/help_text 2019-04-02 11:20:15 -04:00
Ryan Petrello
9036ba492c switch CredentialInput creation to use the associate/disassociate view 2019-04-02 11:20:10 -04:00
Jake McDermott
d87144c4a7 add api for managing credential input sources 2019-04-02 11:19:56 -04:00
Jake McDermott
c209955400 add credential plugin system and minimal working hashivault 2019-04-02 11:19:26 -04:00
Keith Grant
e5f48bfa62 set focus on first input in modal when launching jt 2019-04-02 11:10:02 -04:00
softwarefactory-project-zuul[bot]
6e2c04e16c Merge pull request #3579 from ansible/revert-3575-update_preload_data
Revert "Add scm_branch while preloading data"

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-02 13:02:19 +00:00
softwarefactory-project-zuul[bot]
8f096d11b6 Merge pull request #3569 from ansible/inv_src_venv
Use inventory source model, as opposed to organization, for update custom virtualenv

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-02 12:16:50 +00:00
softwarefactory-project-zuul[bot]
b98bac0ee9 Merge pull request #3580 from ryanpetrello/py3-insights-error
fix a py3 compat issue in the insights action plugin

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-01 21:35:07 +00:00
Ryan Petrello
52a2ed2786 fix a py3 compat issue in the insights action plugin 2019-04-01 17:07:37 -04:00
Vismay Golwala
9710015a2f Revert "Add scm_branch while preloading data" 2019-04-01 16:56:19 -04:00
softwarefactory-project-zuul[bot]
b70c354dfc Merge pull request #3575 from vismay-golwala/update_preload_data
Add scm_branch while preloading data

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-01 20:39:46 +00:00
Keith Grant
e943ae59b7 auto-focus first input in jt launch modal 2019-04-01 15:49:41 -04:00
softwarefactory-project-zuul[bot]
6e8809fe72 Merge pull request #3528 from elyezer/e2e-users
Add auditor and admin params to user e2e

Reviewed-by: Elyézer Rezende
             https://github.com/elyezer
2019-04-01 19:30:39 +00:00
Elyézer Rezende
d2d1074e8b Add auditor and admin params to user e2e 2019-04-01 14:48:33 -04:00
softwarefactory-project-zuul[bot]
1bcda0a4cb Merge pull request #3529 from chrismeyersfsu/fix-inventory_update_deadlock
fixes inventory update deadlock

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-01 18:05:36 +00:00
Marliana Lara
4e4afb4a98 Add sorting to all occurrences of the Template list 2019-04-01 13:44:24 -04:00
AlanCoding
750c8d8a6d Set ansible-inventory verbosity for Ansible 2.8 2019-04-01 13:39:20 -04:00
Vismay Golwala
74e264a6a5 Add scm_branch while preloading data
Currently, the SCM url for default project is not able to
fetch playbooks as branch is not specified. So, adding the
scm_branch value resolves this issue.

Signed-off-by: Vismay Golwala <vgolwala@redhat.com>
2019-04-01 11:53:34 -04:00
softwarefactory-project-zuul[bot]
c508dc7d83 Merge pull request #3564 from wenottingham/paranormal-activity-stream
Management jobs don't produce activity stream entries; don't try and show them

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-04-01 15:23:35 +00:00
AlanCoding
ea355f6f8f change default to allow empty groups from custom sources 2019-04-01 10:53:06 -04:00
Keith Grant
95f5315fa3 prevent console errors when code mirror parses invalid syntax 2019-04-01 08:16:19 -04:00
mabashian
e99e97bcb8 Adds support for custom_virutalenv on inventory sources in the UI 2019-03-31 13:45:24 -04:00
softwarefactory-project-zuul[bot]
7b16931658 Merge pull request #2736 from AlanCoding/logstash_removal
Remove python-logstash dependency

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-31 02:35:45 +00:00
softwarefactory-project-zuul[bot]
e395b87a1b Merge pull request #3559 from chrismeyersfsu/fix-isolated_fact_cache
process host facts for iso runs

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-29 20:44:04 +00:00
chris meyers
71fcb1a82c process host facts for iso runs
* Move isolated clean to our final run hook
* ISO and non-iso code path now share the post-fact-processing code
2019-03-29 16:16:22 -04:00
softwarefactory-project-zuul[bot]
21e5179a84 Merge pull request #3563 from shanemcd/secret-secret
Move secret key from configmap to secret

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-29 19:47:26 +00:00
Shane McDonald
298eaa0b32 Move secret key from configmap to secret 2019-03-29 15:24:50 -04:00
Bill Nottingham
17c0293065 Management jobs don't produce activity stream entries; don't try and show them. 2019-03-29 15:21:22 -04:00
AlanCoding
6f150f0362 Set inv src custom virtualenv directly on model 2019-03-29 14:37:28 -04:00
softwarefactory-project-zuul[bot]
71718ee2eb Merge pull request #3558 from ryanpetrello/index-job-status
index unified_job.status

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-29 18:14:53 +00:00
Ryan Petrello
7ef68a03c6 index unified_job.status
this is a fairly importantj column that we join on for a number of
queries throughout AWX, particularly the task scheduler and analytics
2019-03-29 12:59:43 -04:00
softwarefactory-project-zuul[bot]
bf3473d394 Merge pull request #3552 from ryanpetrello/final-isolated-cleanup
move awx.main.expect to awx.main.isolated

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-29 16:57:02 +00:00
softwarefactory-project-zuul[bot]
eab9ee5128 Merge pull request #3514 from AlanCoding/azure_huzzah
Enable azure_rm inventory plugin

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-29 16:41:16 +00:00
Ryan Petrello
563a0cc2a4 move awx.main.expect to awx.main.isolated 2019-03-29 12:14:40 -04:00
softwarefactory-project-zuul[bot]
3007b9c66a Merge pull request #3518 from jbradberry/isolated-heartbeat
Make use of user-defined settings for defining the isolated nodes heartbeat

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-29 15:15:49 +00:00
softwarefactory-project-zuul[bot]
a4ec149344 Merge pull request #3555 from ryanpetrello/even-more-iso-fixes
fix a variety of bugs in isolated support

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-29 15:13:23 +00:00
Ryan Petrello
a59bc33280 slightly refactor isolated event consumption 2019-03-29 10:48:42 -04:00
Ryan Petrello
8f089c02a5 fix some faulty logic in isolated syncs that caused SSH keys to not work 2019-03-29 10:32:24 -04:00
Ryan Petrello
d663d397f8 clean up some isolated adhoc code
if it's adhoc, there's not project directory to copy
2019-03-29 10:31:56 -04:00
Ryan Petrello
546281d435 work around a bug where runner doesn't provide atomic event writes 2019-03-29 10:31:12 -04:00
softwarefactory-project-zuul[bot]
4e01b11577 Merge pull request #3503 from kialam/fix-3421-wf-viz-popover
Format jt details for popover use.

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-29 13:59:07 +00:00
Kia Lam
d25f1d1357 Format jt details for popover use. 2019-03-29 09:22:15 -04:00
softwarefactory-project-zuul[bot]
2ab290ff2d Merge pull request #3534 from ryanpetrello/iso-pexpect-cleanup
replace our usage of pexpect in IsolatedManager with ansible-runner

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-29 13:17:05 +00:00
softwarefactory-project-zuul[bot]
ef88507d23 Merge pull request #3543 from AlanCoding/do_not_transform
Set Ansible name transform setting for built-in sources

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-28 23:47:56 +00:00
softwarefactory-project-zuul[bot]
df9a012013 Merge pull request #3284 from ansible/analytics
Analytics

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-28 22:40:18 +00:00
Ryan Petrello
9160d91278 clean up old usage of idle_timeout
cmeyers and I looked at this and can't tell where/why/how you'd actually
set this setting - it looks like really old ~2014-2015 Tower history
that probably isn't actually in use
2019-03-28 17:31:32 -04:00
Ryan Petrello
ab11f18957 send an EOF event if isolated dispatch fails 2019-03-28 16:48:52 -04:00
Ryan Petrello
ea30547754 remove main.expect tests (this functionality exists in runner now) 2019-03-28 16:13:14 -04:00
Christian Adams
4a0778a3d5 exclude implicit project updates 2019-03-28 16:12:30 -04:00
softwarefactory-project-zuul[bot]
1270949909 Merge pull request #3539 from AlanCoding/gce_huzzah
Use option to add back in gce_image hostvar

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-28 18:37:59 +00:00
AlanCoding
180396b5df Use option to add back in gce_image hostvar
construct groups from this gce image hostvar
2019-03-28 14:14:44 -04:00
softwarefactory-project-zuul[bot]
504b17b474 Merge pull request #3540 from ansible/ci-repairs
ci adjustments

Reviewed-by: Jake McDermott <yo@jakemcdermott.me>
             https://github.com/jakemcdermott
2019-03-28 17:52:02 +00:00
Ryan Petrello
dd81f59d9a set PYTHONPATH properly for isolated usage of runner 2019-03-28 12:40:21 -04:00
Ryan Petrello
99478f5d25 replace our usage of pexpect in IsolatedManager with ansible-runner 2019-03-28 12:40:21 -04:00
John Mitchell
b3ad12f31a move wf and jt form launch buttons to the left instead of the right 2019-03-28 11:32:12 -04:00
chris meyers
926d6a6525 include tower header to insights on plan fetch
* We include a special header value in the user agent when tower proxies
requests to get per-host rules.
* This extends that header logic to when we fetch plans (playbooks)
2019-03-28 10:56:56 -04:00
AlanCoding
e84642b4a1 Set Ansible name transform setting for built-in sources
This is the current default but will change in the future
for these sources, we do not want behavior change
2019-03-28 10:37:36 -04:00
Daniel Sami
ef3c0cfb38 ci adjustments
lint
2019-03-28 09:52:43 -04:00
Christian Adams
9a4439e731 collect jt table for analytics 2019-03-28 00:58:18 -04:00
softwarefactory-project-zuul[bot]
9479b1b824 Merge pull request #3535 from shanemcd/oops
Fix bug where init scripts didnt create the admin user correctly

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-28 00:39:05 +00:00
Christian Adams
c481919a49 check for open license 2019-03-27 19:54:25 -04:00
Christian Adams
58f0e42bd6 update setting wording & add instance-enabled field 2019-03-27 19:54:25 -04:00
John Mitchell
70af2dd66b add ui for insights enablement 2019-03-27 19:54:25 -04:00
Christian Adams
40dbe70854 update setting wording & add instance-enabled field 2019-03-27 19:53:00 -04:00
Christian Adams
a7368cec43 add copy events table and remove events queries
add copy unifiedjob table, futher optimization, rm since from job_instance queries
2019-03-27 19:53:00 -04:00
Christian Adams
636153d92c add insights setting, optimize and consolidate queries 2019-03-27 19:53:00 -04:00
Christian Adams
ae9377e0e4 add basic job & JT analytics 2019-03-27 19:53:00 -04:00
Ryan Petrello
c586fa9821 add a minimal framework for generating analytics/metrics
annotate queries & add license analytics
2019-03-27 19:53:00 -04:00
Shane McDonald
fcf6b4ae45 Fix bug where init scripts didnt create the admin user correctly 2019-03-27 19:43:47 -04:00
softwarefactory-project-zuul[bot]
7b4c63037a Merge pull request #3523 from ryanpetrello/iso-cancel
properly handle isolated cancellation

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-27 22:44:10 +00:00
softwarefactory-project-zuul[bot]
37a90320ec Merge pull request #3532 from jbradberry/double-jeopardy
Log errors directly from inventory_import.py only if running by hand

Reviewed-by: awxbot
             https://github.com/awxbot
2019-03-27 22:27:58 +00:00
Jeff Bradberry
a803e86a95 Log errors directly from inventory_import.py only if running by hand 2019-03-27 18:02:46 -04:00
chris meyers
cb4d55b47a fixes inventory update deadlock
* all inventory updates continue to occur in parallel up to the point
that they update the database with their results.
* the "funnel" is achieved by using a global per-inventory postgres
named lock
2019-03-27 16:52:07 -04:00
Jeff Bradberry
e398a0ac5e Remove the isolated_heartbeat setup from settings/defaults.py 2019-03-27 14:12:42 -04:00
softwarefactory-project-zuul[bot]
196a6ff36c Merge pull request #3525 from shanemcd/make-things-work
Fix docker-compose installs

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-27 17:38:36 +00:00
Shane McDonald
c3ba851908 Fix docker-compose installs
In a series of unfortunate events, my patch yesterday didnt actually work. This fixes that.
2019-03-27 13:06:55 -04:00
softwarefactory-project-zuul[bot]
11223472d3 Merge pull request #3524 from kdelee/update_install_docs
update install docs

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-27 16:26:03 +00:00
softwarefactory-project-zuul[bot]
d0a996b139 Merge pull request #3520 from ryanpetrello/runner131
pin runner 1.3.1

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-27 15:58:34 +00:00
Elijah DeLee
7dd635cd8d update install docs 2019-03-27 11:54:33 -04:00
softwarefactory-project-zuul[bot]
f715e4e410 Merge pull request #3519 from chrismeyersfsu/fix-iso_bwrap
runner expects process isolation flags in settings

Reviewed-by: awxbot
             https://github.com/awxbot
2019-03-27 15:49:24 +00:00
Ryan Petrello
a983d4bc1f properly handle isolated cancellation 2019-03-27 11:44:26 -04:00
Ryan Petrello
f7cffbfe5c pin runner 1.3.1 2019-03-27 11:23:25 -04:00
chris meyers
2329079326 runner expects process isolation flags in settings
* Towards the goal of converging the iso code path w/ the non-iso code
path. More process isolation control flags into settings.
2019-03-27 11:08:41 -04:00
Jeff Bradberry
bc6df9cab8 Make use of user-defined settings for defining the isolated nodes heartbeat
Just in case the user has created custom settings files which set
AWX_ISOLATED_PERIODIC_CHECK.
2019-03-27 10:47:38 -04:00
softwarefactory-project-zuul[bot]
055e7b4974 Merge pull request #3515 from shanemcd/docker-compose-permissions
Fix permissions of sensitive files in docker-compose installation

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-27 14:11:31 +00:00
Jeff Bradberry
efb4fb6fd0 Remove some no longer used imports 2019-03-27 10:06:13 -04:00
Jeff Bradberry
f2be4de544 Use Django's own logic to invalidate sessions of users when changing passwords
The key is django.contrib.auth.update_session_auth_hash(), which knows
how to inject a recalculated session hash back into the session if the
requesting user is changing their own password, in order to keep that
user logged in.
2019-03-27 10:06:13 -04:00
Shane McDonald
c44bf6f903 Allow for platform specific variables in docker-compose install
This changes the default docker_compose_dir on macos to a writeable location
2019-03-27 09:32:04 -04:00
Shane McDonald
a6d031f46f Fix permissions of sensitive files in docker-compose installation 2019-03-27 09:31:10 -04:00
AlanCoding
73f16b2bee Enable azure_rm inventory plugin 2019-03-27 09:27:10 -04:00
softwarefactory-project-zuul[bot]
2129f12085 Merge pull request #3505 from shanemcd/devel
Address CVE-2019-3869

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-26 19:38:28 +00:00
softwarefactory-project-zuul[bot]
23185ca22f Merge pull request #3232 from AlanCoding/truly_empty_groups
Surface empty groups as children of all group

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-26 19:17:36 +00:00
Shane McDonald
2b6cf97157 Do not set credentials via environment variables 2019-03-26 15:13:28 -04:00
Shane McDonald
07e5a00f14 Remove “standalone Docker” installation path
This has been a burden to maintain. docker-compose is now required
2019-03-26 15:13:28 -04:00
softwarefactory-project-zuul[bot]
1b0f5b05ad Merge pull request #3502 from marshmalien/feat-toolbar-sort-instance-modal-list
Feat toolbar sort instance modal list

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-26 18:57:24 +00:00
softwarefactory-project-zuul[bot]
5ff4625eb1 Merge pull request #3280 from AlanCoding/playbook_dir
Set --playbook-dir in calls to ansible-inventory

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-26 18:53:12 +00:00
Marliana Lara
1829e7cad4 Add sort toolbar to instance modal 2019-03-26 14:34:33 -04:00
AlanCoding
e097f5a021 implement playbook-dir option in ansible-inventory calls 2019-03-26 14:09:08 -04:00
softwarefactory-project-zuul[bot]
caa5596386 Merge pull request #3320 from vismay-golwala/custom_venvs
Feature: custom virtual environment directories

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-26 16:26:50 +00:00
Vismay Golwala
ec390b049d Feature: custom virtual environment directories
Currently, users are allowed to define virtual environments in
`settings.BASE_VENV_PATH` only, because that's the only place
Tower looks for virtual environments. This feature allows users
to custom define the directory paths, using API or UI, to look
for virtual environments. Tower aggregates virtual environments
from all these paths, except environments with special name `awx`.

Signed-off-by: Vismay Golwala <vgolwala@redhat.com>
2019-03-26 12:04:17 -04:00
softwarefactory-project-zuul[bot]
0814a9c4a1 Merge pull request #3266 from ansible/inventory_plugins
Transition to inventory plugins

Reviewed-by: Alan Rominger <arominge@redhat.com>
             https://github.com/AlanCoding
2019-03-26 15:14:27 +00:00
Marliana Lara
0a1b220f56 Link to instances list via sref directive 2019-03-26 10:35:22 -04:00
AlanCoding
d39b3b3165 Remove compatibility_mode field, simplify jinja2 syntax
fix minor bug where credential not shown in API
2019-03-26 10:29:39 -04:00
AlanCoding
19ad7d3983 Inventory plugins data tweaks and finalization
Disable use of azure_rm inventory plugin
Disable use of ec2 inventory plugin
due to compatibility issues that are unresolved

Fix conflicts with ansible runner integration

Add additional content enabled by Ansible core changes
2019-03-26 10:29:39 -04:00
AlanCoding
cd7e358b73 Inventory plugins transition dev finishing work
Bump keystone auth to resolve problem with openstack script

Clarify code path, routing to template vs. managed injector
  behavior is also now reflected in test data files

Refactor test data layout for inventory injector logic

Add developer docs for inventory plugins transition

Memoize only get_ansible_version with no parameters

Make inventory plugin injector enablement a separate
  concept from the initial_version
  switch tests to look for plugin_name as well

Add plugin injectors for tower and foreman.

Add jinja2 native types compat feature

move tower source license compare logic to management command

introduce inventory source compat mode

pin jinja2 for native Ansible types

Add parent group keys, and additional translations

manual dash sanitization for un-region-like ec2 groups

nest zones under regions using Ansible core feature just merged
  implement conditionally only with BOTH group_by options

Make compat mode default be true
  in API models, UI add and edit controllers

Add several additional hostvars to translation
Add Azure tags null case translation

Make Azure group_by key off source_vars
  to be consistent with the script

support top-level ec2 boto_profile setting
2019-03-26 10:29:39 -04:00
AlanCoding
bc5881ad21 Primary development of inventory plugins, partial compat layer
Initialize some inventory plugin test data files
Implement openstack inventory plugin

This may be removed later:
- port non-JSON line strip method from core

Dupliate effort with AWX mainline devel
- Produce ansible_version related to venv

Refactor some of injector management, moving more
  of this overhead into tasks.py, when it comes to
  managing injector kwargs

Upgrade and move openstack inventory script
  sync up parameters

Add extremely detailed logic to inventory file creation
for ec2, Azure, and gce so that they are closer to a
genuine superset of what the contrib script used to give.
2019-03-26 10:29:39 -04:00
Jim Ladd
dd854baba2 Add support for Azure inventory plugin 2019-03-26 10:29:39 -04:00
Jim Ladd
7cce3cad06 Add support for ec2 inventory plugin 2019-03-26 10:29:38 -04:00
AlanCoding
622fbc116b move script injection logic to inventory file 2019-03-26 10:29:38 -04:00
AlanCoding
b9d489c788 Use randomized file names for injector credential files 2019-03-26 10:29:38 -04:00
AlanCoding
5cbcfbe0c6 Port inventory source injector tests to functional tests
This new batch of tests assures that the injector logic
for inventory source in their old script version remains
untouched with the refactoring underway.

Plugins are also tested by the same means of comparing
to reference files, these will be used to assure that
all parameters that used to be respected are still
respected in the plugin system.
2019-03-26 10:29:38 -04:00
Jim Ladd
d46a403a49 GCE plugin should not set any regions when 'all' specified 2019-03-26 10:29:38 -04:00
Jim Ladd
de808d4911 Only install futures on py2 2019-03-26 10:29:38 -04:00
AlanCoding
43eff55fd4 fix bugs related to python3 2019-03-26 10:29:37 -04:00
AlanCoding
6c130fa6c3 Build-in inventory plugin code structure with gce working
supporting and related changes
 - Fix inconsistency between can_update / can_start
 - Avoid creating inventory file twice unnecessarily
 - Non-functional consolidation in Azure injection logic
 - Inject GCE creds as indented JSON for readability
 - Create new injector class structure, add gce
 - Reduce management command overrides of runtime environment
2019-03-26 10:29:35 -04:00
softwarefactory-project-zuul[bot]
90ea9a8cc4 Merge pull request #3500 from Spredzy/etc_ssh
bwrap/runner: Add /etc/ssh in bind mounted folder

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-26 12:27:43 +00:00
softwarefactory-project-zuul[bot]
b09bca54b7 Merge pull request #3499 from rooftopcellist/content-oauth
update content-type for oauth2 docs

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-26 12:07:32 +00:00
Ryan Petrello
8e4a87d0af fix tests for add /etc/ssh in bind mounted folder 2019-03-26 08:04:16 -04:00
Yanis Guenane
fd50feb258 bwrap/runner: Add /etc/ssh in bind mounted folder
/etc/ssh is currently not bound when run into bwrap, this leads to
error like "Bad owner or permissions on /etc/ssh/ssh_config.d/05-redhat.conf"
since it cannot access this file.

https://github.com/ansible/awx/pull/3391 was done pre runner
integration.

Fixes: https://github.com/ansible/awx/issues/3392

Signed-off-by: Yanis Guenane <yanis@guenane.org>
2019-03-26 12:43:59 +01:00
AlanCoding
f749a5d44d Surface empty groups as children of all group 2019-03-26 07:18:52 -04:00
Christian Adams
c3366db5ca update content-type for oauth2 docs 2019-03-25 23:55:56 -04:00
softwarefactory-project-zuul[bot]
07a9cd106e Merge pull request #3494 from ryanpetrello/more-iso
more iso cleanup and bug fixes

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-25 23:21:06 +00:00
Ryan Petrello
b2a1824d21 store set_stat data for isolated job runs 2019-03-25 18:53:42 -04:00
softwarefactory-project-zuul[bot]
303443796e Merge pull request #3458 from marshmalien/feat-toolbar-sort-user-tokens
Add sort to User Tokens list

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-25 22:25:55 +00:00
Ryan Petrello
495dc2202f more iso cleanup and bug fixes 2019-03-25 17:47:58 -04:00
softwarefactory-project-zuul[bot]
33f5200a20 Merge pull request #3497 from ryanpetrello/whoops-dot-js
remove an errant console.log

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-25 19:59:47 +00:00
Ryan Petrello
8674e3b4de remove an errant console.log 2019-03-25 15:24:34 -04:00
softwarefactory-project-zuul[bot]
ace459cf70 Merge pull request #3447 from beeankha/node_activity_stream
WFJT Node Activity Stream Bug Fix

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-25 19:23:27 +00:00
softwarefactory-project-zuul[bot]
d0c952692d Merge pull request #3481 from ryanpetrello/minor-runner-cleanup
Minor post-runner cleanup

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-25 16:57:15 +00:00
Ryan Petrello
af8e071840 remove old callback plugin code and tests 2019-03-25 12:26:51 -04:00
Ryan Petrello
e6abd77c96 remove py2 compatability from awx.main.expect.run
this module is no longer run on isolated nodes (they use runner now)
2019-03-25 12:26:47 -04:00
Ryan Petrello
42bfff301c remove python-memcached as a base dependency for playbook execution 2019-03-25 12:26:44 -04:00
Marliana Lara
0aff1a2c75 Add sort to users tokens list 2019-03-25 11:18:18 -04:00
Keith Grant
685f4018f2 improve verbiage in activity stream for associating/disassociating wf nodes 2019-03-25 10:59:24 -04:00
softwarefactory-project-zuul[bot]
1dff691830 Merge pull request #3468 from kialam/fix-groups-list-view
Fix 'Groups' list item styling.

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-25 14:54:25 +00:00
softwarefactory-project-zuul[bot]
525021214c Merge pull request #3483 from keithjgrant/inventory-vars-popout
Inventory vars popout round 2 (codemirror)

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-22 20:27:13 +00:00
softwarefactory-project-zuul[bot]
c12c64f5e7 Merge pull request #3484 from shanemcd/devel
Fix python3 offline installs

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-22 18:31:05 +00:00
Shane McDonald
0eaeadad87 Fix python3 offline installs 2019-03-22 13:30:09 -04:00
softwarefactory-project-zuul[bot]
eb5846d1be Merge pull request #3482 from chrismeyersfsu/fix-iso_ident
do not generate a random ident

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-22 16:43:04 +00:00
chris meyers
87e1ba4dea do not generate a random ident
* instead, set the ident passed to ansible runner to be the job id. That
way, on we know what directory to look in for results when the directory
structure is created.
2019-03-22 12:19:42 -04:00
softwarefactory-project-zuul[bot]
a9427dbf1b Merge pull request #3459 from beeankha/email_timeout_option
Provide Default Email Timeout Value

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-22 15:37:02 +00:00
Keith Grant
e96e1e925c update e2e test for codemirror change 2019-03-22 11:30:31 -04:00
Keith Grant
7476fefd65 fix codemirror in add host, add inventory, and workflow job details 2019-03-22 10:11:18 -04:00
softwarefactory-project-zuul[bot]
8b2fc26219 Merge pull request #3041 from chrismeyersfsu/runnerpy3
ansible-runner integration

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-21 22:21:50 +00:00
softwarefactory-project-zuul[bot]
9480f911b2 Merge pull request #3471 from ansible/add_dev_supervisor
Install supervisor into the dev environment

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-21 20:39:03 +00:00
beeankha
d7fc3f53b8 Update indentation 2019-03-21 16:14:56 -04:00
Matthew Jones
91cbaa1096 Install supervisor into the dev environment 2019-03-21 15:58:24 -04:00
beeankha
3e13eff7f4 Change serializer to take in init paramdefault value 2019-03-21 15:22:34 -04:00
softwarefactory-project-zuul[bot]
a562994b64 Merge pull request #3466 from ryanpetrello/libcloud-minus-pycrypto
pin apache-libcloud to a version that doesn't use PyCrypto

Reviewed-by: Matthew Jones <mat@matburt.net>
             https://github.com/matburt
2019-03-21 18:59:28 +00:00
Ryan Petrello
b02d9ae282 pin apache-libcloud to a version that doesn't use PyCrypto
once a new version lands on PyPI, we'll pin to _it_
2019-03-21 14:21:04 -04:00
Kia Lam
57820b7056 Fix 'Groups' list item styling. 2019-03-21 13:44:56 -04:00
softwarefactory-project-zuul[bot]
e3bbd436b4 Merge pull request #3215 from vismay-golwala/survey_allow_empty_defaults
Allow empty default values for numerical survey answers.

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-21 16:37:02 +00:00
softwarefactory-project-zuul[bot]
9aa9524257 Merge pull request #3440 from marshmalien/feat-toolbar-sort-instances-list
Add sort and pagination to instances list

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-21 16:19:50 +00:00
softwarefactory-project-zuul[bot]
af5a898919 Merge pull request #3433 from jlmitch5/lookupEnhancement
hide rows from lists where adding would be redundant

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-21 15:09:17 +00:00
softwarefactory-project-zuul[bot]
a04329efed Merge pull request #3453 from marshmalien/feat-toolbar-sort-application-tokens-list
Add sort to application tokens list

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-21 14:58:52 +00:00
Keith Grant
a0b2ce3ef1 Merge pull request #3451 from keithjgrant/inventory-vars-popout
Host/Inventory vars popout (code mirror)
2019-03-21 10:55:12 -04:00
mabashian
b394766075 Show manual project path when available 2019-03-21 10:45:16 -04:00
softwarefactory-project-zuul[bot]
cd62f39bce Merge pull request #3219 from mickfeech/devel
update documentation to include kuberentes initContainers

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-21 14:36:17 +00:00
chris meyers
b7b97dd58d doc update fix 2019-03-21 09:26:59 -04:00
beeankha
1d03625b27 Remove comment from serializer 2019-03-21 09:08:36 -04:00
softwarefactory-project-zuul[bot]
af55c4c05e Merge pull request #3455 from ansible/websockets-again
e2e fixtures and WS

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-21 12:51:26 +00:00
chris meyers
0a670e8db1 change from runner master to runner 1.3 2019-03-21 07:46:11 -04:00
Keith Grant
cf62fa67bd add links to activity stream for workflow_job_template_node 2019-03-20 17:00:42 -04:00
mickfeech
3c382322b0 Fix misspelled word 2019-03-20 16:33:02 -04:00
mickfeech
f4ef3024fd Merge branch 'devel' of https://github.com/mickfeech/awx into devel 2019-03-20 16:30:55 -04:00
Unknown
67ca2fa335 update documentation to include kuberentes initContainers
Update documentation to include Kubernetes initContainers in custom virtualenvs
2019-03-20 16:25:49 -04:00
beeankha
c9ac805eed [WIP] Provide Default Email Timeout Value 2019-03-20 16:17:48 -04:00
softwarefactory-project-zuul[bot]
f40b637efc Merge pull request #3457 from AlexSCorey/3300-JTCreationPermissionOnDashboard
Shows the button the add a JT to users with permissions to make JTs.

Reviewed-by: Alex Corey <Alex.swansboro@gmail.com>
             https://github.com/AlexSCorey
2019-03-20 20:16:37 +00:00
chris meyers
60ef160e85 flake8 fix 2019-03-20 16:12:45 -04:00
Vismay Golwala
df11a7fd3d Project admin manual SCM Type creation bug fix
There was a bug in tower where project admin was not allowed to create
a project with manual SCM type. This was because, the project base dir
was only set from `settings.PROJECTS_ROOT`, if the user role was either
admin or auditor. This has been updated by also allowing it to set when
user is a project admin.

Signed-off-by: Vismay Golwala <vgolwala@redhat.com>
2019-03-20 16:01:33 -04:00
AlanCoding
1ae4ed4922 migrate python-logstash license 2019-03-20 15:29:06 -04:00
AlanCoding
8a72a4d39d Prune the python2 specific logic from log formatter 2019-03-20 15:29:06 -04:00
AlanCoding
127495b53d remove things from base class that were never used 2019-03-20 15:29:05 -04:00
AlanCoding
017d367749 Remove dependency and insert class 2019-03-20 15:29:05 -04:00
Alex Corey
be507dbefb Shows the button the add a JT to users with permissions to make JTs. 2019-03-20 14:55:28 -04:00
chris meyers
8c26f20188 add license files for python modules
* python-daemon
* ansible-runner
2019-03-20 14:51:41 -04:00
chris meyers
2c52a7d9a8 fix more unit tests for runner
* isolated will be fixed in the future so pytest skip those
* fact cache moved one directory level up, account for that
2019-03-20 14:32:52 -04:00
chris meyers
b006510035 do not save sensitive env vars
* job_env gets exposed via the api. Sensitive env variables should be
redacted before saved into job_env.
2019-03-20 14:00:22 -04:00
Keith Grant
8e48a3a523 interpret empty codemirror json content as empty object 2019-03-20 13:51:10 -04:00
Daniel Sami
b26c8f6b62 e2e adjustments for fixtures and WS
lint adjustment
2019-03-20 12:55:37 -04:00
softwarefactory-project-zuul[bot]
68d7532d01 Merge pull request #3452 from wenottingham/where-did-you-come-from
Add originating address for the failed login message

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-20 16:18:42 +00:00
Keith Grant
e9cf1475ca de-lint 2019-03-20 11:45:51 -04:00
Bill Nottingham
1b3ae50076 Add originating address for the failed login message 2019-03-20 11:34:35 -04:00
John Mitchell
7791c5f5ba hide groups that host is already associated with on relevant lists 2019-03-20 11:34:05 -04:00
Marliana Lara
19abd24c91 Add sort to application tokens list 2019-03-20 11:30:56 -04:00
softwarefactory-project-zuul[bot]
b2ae68850c Merge pull request #3449 from elyezer/users-e2e
users e2e

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-20 15:28:57 +00:00
chris meyers
1a6ae6e107 allow for runner setting parent_uuid
* Previously, parent_uuid was expected only on events generated for a
Job run. Now, there maybe a parent_uuid for any job type. AWX does not
support parenting events for any job type other than Job.
2019-03-20 11:05:01 -04:00
Keith Grant
86c7fd3b5d codemirror: sync data when closing modal by clicking outside 2019-03-20 11:01:42 -04:00
Elyézer Rezende
46ad3fa7b1 users e2e 2019-03-20 09:48:25 -04:00
chris meyers
060585434a update tests 2019-03-20 09:44:38 -04:00
Keith Grant
2657779eda cleanup comments 2019-03-20 08:56:42 -04:00
Keith Grant
ac890b8cda don't show code-mirror tooltip icon if no tooltip provided 2019-03-20 08:36:53 -04:00
Keith Grant
b6d8f9c6f6 use code-mirror directive for host facts 2019-03-20 08:36:52 -04:00
Keith Grant
5583af2a58 form generator: add ng-disabled support to code mirror fields 2019-03-20 08:36:52 -04:00
Keith Grant
2ee6713050 centralize variable parsing logic in code mirror directive 2019-03-20 08:36:52 -04:00
Keith Grant
b28409c1c7 fix saving inventory variables 2019-03-20 08:36:52 -04:00
Keith Grant
ac5dec272b code-mirror: keep yaml/json in sync when opening modal 2019-03-20 08:36:52 -04:00
Keith Grant
33b19ebe1f code-mirror: keep yaml/json setting in sync with modal 2019-03-20 08:36:52 -04:00
Keith Grant
5d3e39beac add use code-mirror directive for host variables; fix multiple code-mirrors on page at once 2019-03-20 08:36:52 -04:00
mabashian
bed63b3690 Work on getting extra vars popout working on inv form 2019-03-20 08:36:52 -04:00
Keith Grant
43ef4183df trying to make codemirror editable 2019-03-20 08:36:52 -04:00
softwarefactory-project-zuul[bot]
74869494f9 Merge pull request #3441 from AlexSCorey/3258-workflowDisplay
makes the card for the workflow display taller

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-19 19:25:27 +00:00
Alex Corey
bd4337976e makes the card for the workflow display taller 2019-03-19 15:02:05 -04:00
John Mitchell
50079c0441 fix issue where pagination would lose role filter 2019-03-19 12:52:59 -04:00
John Mitchell
f3173dbe26 address pr and product feedback 2019-03-19 12:08:36 -04:00
beeankha
a1dd5a4e19 WIP WFJT Node Activity Stream Bug Fix 2019-03-19 11:29:19 -04:00
John Mitchell
52e86cf0c3 hide rows from lists where adding would be redundant 2019-03-19 10:30:31 -04:00
softwarefactory-project-zuul[bot]
3d9a47f0d9 Merge pull request #3424 from falencastro/devel
Makes daphne websocket_timeout infinite.

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-19 12:08:50 +00:00
chris meyers
5135b8a969 fixup unit tests for tasks 2019-03-18 14:21:47 -04:00
chris meyers
8a04c22b2b point at another runner branch
* revert parent_uuid because it causes problems with unexpected
parameter on event creation for some event type.
2019-03-18 14:21:47 -04:00
chris meyers
f7842cf283 refactor and fix unit tests
* fixup task TestGenericRun
* make runner callback functions accessable to testing
* reduce isintance() usage in run() by using build_ pattern
* move process_isolation param building to build_ function so it can be
tested
2019-03-18 14:21:47 -04:00
chris meyers
827ad0fa75 remove safe_args and add status_handler
* safe_args no longer makes sense. We have moved extra_vars to a file
and thus do not pass sensitive content on the cmdline
2019-03-18 14:21:47 -04:00
Ryan Petrello
602ef9750f update isolated task execution for ansible-runner 2019-03-18 14:21:47 -04:00
chris meyers
8fb65b40de use ansible runner to run playbooks
* Project Updates
* Jobs
* Inventory Updates
* System Jobs
* AdHoc Commands

* Notifications
* Fact Cache
* proot
2019-03-18 14:21:47 -04:00
chris meyers
a7cda95803 init ansible-runner requirements 2019-03-18 14:21:47 -04:00
Marliana Lara
bb33ed6415 Add sort and pagination to instances list 2019-03-18 13:36:58 -04:00
softwarefactory-project-zuul[bot]
358ad05e51 Merge pull request #3439 from beeankha/notification_doc
Make edit to Notification doc

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-18 15:54:25 +00:00
beeankha
74c84bd7df Update email section of Notification doc 2019-03-18 11:28:57 -04:00
softwarefactory-project-zuul[bot]
62e1d5fdd2 Merge pull request #3435 from AlexSCorey/3421-fixInventoryGroupColumnV2
Fix inventory column  width issues to make their contents readable

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-18 14:26:17 +00:00
Alex Corey
b8beb1c64e Fix invetory column width issues to make their contents readable 2019-03-18 09:57:42 -04:00
Shane McDonald
14d86ef5d3 Merge pull request #3426 from sugitk/patch-1
propose the change translation in Japanese
2019-03-18 09:36:42 -04:00
softwarefactory-project-zuul[bot]
9ab7752d32 Merge pull request #3416 from Spredzy/urllib_parse
plugins/tower.py: Use urllib.parse rather than urlparse

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-18 13:16:29 +00:00
Yanis Guenane
3a4f56bb2b plugins/tower.py: Use urllib.parse rather than urlparse
urlparse does not exist in python3, it has been replaced by urllib.parse

Signed-off-by: Yanis Guenane <yguenane@redhat.com>
2019-03-18 09:43:24 +01:00
softwarefactory-project-zuul[bot]
8f1c20423b Merge pull request #3222 from beeankha/timeout_config
Add Timeout Feature to Email Notification Template

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-16 12:46:17 +00:00
softwarefactory-project-zuul[bot]
6fd5f9c6d8 Merge pull request #3425 from ryanpetrello/dispatcher-stop-log
send callback receiver log lines to the correct logger

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-15 15:20:50 +00:00
softwarefactory-project-zuul[bot]
6b187946fb Merge pull request #3370 from vismay-golwala/scroll_top
UI - scroll to top in pagination

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-15 13:42:04 +00:00
softwarefactory-project-zuul[bot]
d54f633a7b Merge pull request #3401 from marshmalien/feat-toolbar-sort-applications-list
Add sort toolbar to applications list

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-15 13:32:17 +00:00
softwarefactory-project-zuul[bot]
d0571c2cab Merge pull request #3408 from marshmalien/feat-toolbar-sort-instance-groups-list
Add sort toolbar to instance groups list

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-15 13:21:08 +00:00
Ryan Petrello
32ee9838af use the correct logger for the callback receiver
the callback receiver and dispatcher share several modules, so add logic
to use the correct logger
2019-03-15 08:09:47 -04:00
softwarefactory-project-zuul[bot]
c41068edc4 Merge pull request #3420 from baxeno/https_download_postgresql_rpm
docker: yum: use https for postgresql pgdg rpm download

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-15 12:04:21 +00:00
Takashi Sugimura
f2548c5e66 #3415 propose the change translation in Japanese
regarding https://github.com/ansible/awx/issues/3415
2019-03-15 08:27:41 +09:00
beeankha
66a52655df Change email notification success/fail messages and add a timeout feature 2019-03-14 16:37:22 -04:00
Vismay Golwala
32dbe3f86a UI - scroll to top in pagination
Currently in pagination, when we switch from one page to another,
the view is stuck at the bottom and is slightly inconvenient to
scroll all the way to top, in potentially a large list of records.
So in order to prevent that, this commit sets automatic scroll to
top while switching between pages, using jQuery's animate and
scrollTop methods.

Signed-off-by: Vismay Golwala <vgolwala@redhat.com>
2019-03-14 16:35:30 -04:00
softwarefactory-project-zuul[bot]
c6ae7d84a2 Merge pull request #3417 from AlexSCorey/3395-deleteLongGroupName
Fixes issue with inventory group names that are extremely long.

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-14 20:17:49 +00:00
Felipe Alencastro
7d384262e4 Makes daphne websocket_timeout infinite.
Daphne has a default timeout of 86400 seconds, so after 1 day of starting
awx_web container, the stdout stops refreshing automatically on the web UI.
This fixes this issue by making the timeout infinite, so the connection
between nginx and daphne's websocket never closes.
2019-03-14 17:17:09 -03:00
softwarefactory-project-zuul[bot]
64debd7230 Merge pull request #3423 from kdelee/update_git_repo_for_tests
Update reference to test playbooks

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-14 20:13:12 +00:00
Elijah DeLee
d39cfd1778 Update reference to test playbooks 2019-03-14 14:33:16 -04:00
Bruno Thomsen
2e0edcbabd docker: yum: use https for postgresql rpm download.
Signed-off-by: Bruno Thomsen <bruno.thomsen@gmail.com>
2019-03-14 17:14:17 +01:00
Alex Corey
2650cbfc87 Fixes issue with inventory group names that are extremely long. 2019-03-14 11:16:08 -04:00
softwarefactory-project-zuul[bot]
df72a01f27 Merge pull request #3412 from AlanCoding/put_down_the_dispatcher
Run computed fields once for bulk delete requests

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-14 00:05:09 +00:00
AlanCoding
7cf2bc2410 Run computed fields once for bulk delete requests 2019-03-13 15:37:01 -04:00
softwarefactory-project-zuul[bot]
a63a204a21 Merge pull request #3409 from ansible/thedoubl3j-patch-1
typo in inventory

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-13 18:43:59 +00:00
softwarefactory-project-zuul[bot]
928de6127b Merge pull request #3411 from AlexSCorey/3337-brokenDockLink
fixes broken documentation link

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-13 14:47:48 +00:00
Alex Corey
85eca47a93 fixes broken documentation link 2019-03-12 16:37:15 -04:00
softwarefactory-project-zuul[bot]
99c8c4bf2b Merge pull request #3410 from ryanpetrello/fix-swagger-doc-builds
fix a failing test in the unit-test target used to generate swagger docs

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-12 19:23:40 +00:00
Ryan Petrello
50d8eb30e1 fix a failing test in the unit-test target used to generate swagger docs 2019-03-12 14:59:24 -04:00
softwarefactory-project-zuul[bot]
b1d9b14ab1 Merge pull request #3403 from shanemcd/ootpa
Working out some python3 kinks

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-12 16:00:45 +00:00
Marliana Lara
56b3d6c79b Add sort toolbar to instance groups list 2019-03-12 11:56:26 -04:00
Jake Jackson
9e528ea898 typo in inventory
simple typo fix `this` -> `these`
2019-03-12 11:48:41 -04:00
Shane McDonald
e09684462c Working out some python3 kinks 2019-03-12 11:27:50 -04:00
softwarefactory-project-zuul[bot]
22c4b28917 Merge pull request #3404 from ryanpetrello/fix-missing-swagger-endpoints
generate Swagger schemas as if view permissions didn't matter

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-12 15:09:52 +00:00
softwarefactory-project-zuul[bot]
9d0a8d2047 Merge pull request #3377 from mabashian/root-all-groups-toggle
Adds toggle for all/root groups to inventory groups view

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-12 14:09:45 +00:00
Ryan Petrello
023fbc931d generate Swagger schemas as if view permissions didn't matter
this fixes several scenarios where certain POST endpoints don't show up
in our generated Swagger doc; /api/swagger is only
discoverable/accessible in the development environment where we generate
the schema
2019-03-11 21:26:32 -04:00
softwarefactory-project-zuul[bot]
c3ae700888 Merge pull request #3402 from AlexSCorey/3238-addFirstNameLastNameonUsersTeamList
add first name and last name to the headers row of the teams users list

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-11 20:25:36 +00:00
Marliana Lara
ee6445d620 Update dataset from event listener instead of queryset search method 2019-03-11 15:08:53 -04:00
Alex Corey
8fb7cb6e82 add first name and last name to the headers row of the teams users list 2019-03-11 14:16:26 -04:00
Marliana Lara
b55212368b Add sort toolbar to applications list 2019-03-11 13:44:21 -04:00
softwarefactory-project-zuul[bot]
649d854225 Merge pull request #3394 from jbradberry/update-become-methods-list
Add the ksu, machinectl, and sesu methods to the builtin list of become methods

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-11 15:25:31 +00:00
softwarefactory-project-zuul[bot]
fb1d918c2d Merge pull request #3391 from Spredzy/add_etc_ssh_in_ro_bind
bwrap: Add /etc/ssh in bind mounted folder

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-08 17:46:17 +00:00
Jeff Bradberry
e8d93c99a6 Add the ksu, machinectl, and sesu methods to the builtin list of become methods 2019-03-08 11:14:18 -05:00
softwarefactory-project-zuul[bot]
b54ec6b9c8 Merge pull request #3389 from keithjgrant/launch-tooltips
add tooltips for fields prompted on launch

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-08 16:12:59 +00:00
Keith Grant
3acb474b19 de-dupe element ids 2019-03-08 10:16:42 -05:00
softwarefactory-project-zuul[bot]
2e0d381f8f Merge pull request #3386 from jbradberry/org-limits-message
Update the error message when exceeding the organization hosts limit

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-08 15:09:04 +00:00
Yanis Guenane
7eb483d810 bwrap: Add /etc/ssh in bind mounted folder
/etc/ssh is currently not bound when run into bwrap, this leads to error
like "Bad owner or permissions on /etc/ssh/ssh_config.d/05-redhat.conf"
since it cannot access this file.
2019-03-08 15:20:53 +01:00
softwarefactory-project-zuul[bot]
7b570b59c6 Merge pull request #3367 from marshmalien/feat-toolbar-sort-jobs-list
Add toolbar sort to Jobs lists

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-08 14:18:15 +00:00
Keith Grant
09f9204917 add tooltips for fields prompted on launch 2019-03-07 16:27:01 -05:00
Jeff Bradberry
2a8e6ecba1 Update the error message when exceeding the organization hosts limit 2019-03-07 14:13:54 -05:00
Marliana Lara
abb221d942 Stretch Layout container to full height 2019-03-07 13:44:15 -05:00
softwarefactory-project-zuul[bot]
c8fdf46dda Merge pull request #3360 from jlmitch5/addUnsavedWorkflowChanges
add unsaved workflow changes flow

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-07 18:37:19 +00:00
softwarefactory-project-zuul[bot]
a9226fc25f Merge pull request #3375 from shanemcd/devel
Fix dev environment when running as root on the host

Reviewed-by: Matthew Jones <mat@matburt.net>
             https://github.com/matburt
2019-03-07 17:20:32 +00:00
softwarefactory-project-zuul[bot]
67eba3cf5c Merge pull request #3381 from matburt/revert_xlib_deps
Revert xlib deps

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-07 16:03:33 +00:00
softwarefactory-project-zuul[bot]
20b8cdfb3d Merge pull request #3382 from jakemcdermott/devdocs
update docs for development environment

Reviewed-by: Christian Adams <rooftopcellist@gmail.com>
             https://github.com/rooftopcellist
2019-03-07 15:58:07 +00:00
softwarefactory-project-zuul[bot]
1dcb7591c5 Merge pull request #3363 from mabashian/dashboard-dropdowns
Fix dashboard dropdowns after bootstrap upgrade

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-07 15:38:11 +00:00
Jake McDermott
bca9735534 update docs for development environment 2019-03-07 10:30:31 -05:00
softwarefactory-project-zuul[bot]
f95576764d Merge pull request #3373 from AlexSCorey/2960-addTemplateTitlePromptDiag
add template name to launch prompt modal.

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-07 15:21:45 +00:00
softwarefactory-project-zuul[bot]
92a600aaa9 Merge pull request #3359 from mabashian/yaml-comments
Show yaml comments when possible on launch prompt

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-07 15:14:19 +00:00
Matthew Jones
5cf7cc21c8 Revert "Fix chrome can not be started with unit-tests due to missing shared libraries"
This reverts commit d558ffd699.
2019-03-07 10:07:08 -05:00
Matthew Jones
499fd7b2f1 Revert "Fix chrome can not be started with unit-tests due to missing shared libraries"
This reverts commit 3e5f328b52.
2019-03-07 10:06:30 -05:00
Alex Corey
0593ac197c add template name to launch prompt modal. 2019-03-07 09:42:59 -05:00
softwarefactory-project-zuul[bot]
4fc0d220cc Merge pull request #3376 from keithjgrant/fix-save-confirmation-modal
fix save success modal when adding new application

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-07 14:23:34 +00:00
softwarefactory-project-zuul[bot]
d309acfddb Merge pull request #3372 from beeankha/jt_strings2
Update fields.py to Display Correct Error Message When JT Credential Is Not an Integer

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-06 23:32:38 +00:00
beeankha
2196089216 Update fields.py to convert JT Credential input into integers 2019-03-06 18:05:53 -05:00
mabashian
b6bf68427a Remove unused variable 2019-03-06 17:40:33 -05:00
mabashian
502decf8fe Adds toggle for all/root groups to inventory groups view 2019-03-06 17:31:54 -05:00
Keith Grant
20347420ca fix save success modal when adding new application 2019-03-06 17:22:25 -05:00
Shane McDonald
b29a9cd86e Fix dev environment when running as root on the host
Without this, CURRENT_UID isnt actually passed in from the host, and wipes out /etc/passwd even when we’re actually running as root.

I tested this as a non-root user on Linux, and on Docker for Mac
2019-03-06 17:08:56 -05:00
Marliana Lara
2d15d13359 Refactor styles and add queryset to updateDataset event emmitters 2019-03-06 15:09:46 -05:00
Marliana Lara
31f5d13a69 Initialize paginate queryset with an empty object 2019-03-06 12:19:55 -05:00
Marliana Lara
67753b790c Move toolbar default variable above the toolbar setting funciton 2019-03-06 10:48:14 -05:00
Marliana Lara
661a54d356 Add toolbar sort to Jobs list 2019-03-05 15:09:12 -05:00
Marliana Lara
970a714291 Update URL params when sorting on projects list 2019-03-05 15:08:24 -05:00
John Mitchell
dc206c9ad6 fix wf unsaved changes exit functionality 2019-03-05 14:20:38 -05:00
softwarefactory-project-zuul[bot]
658bdddac3 Merge pull request #3354 from ansible/404_test
404 validation

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-05 17:19:52 +00:00
John Mitchell
9f5f86c6a7 remove unused service and fix lint issue 2019-03-05 12:00:12 -05:00
mabashian
5e37882267 Remove errant console 2019-03-05 11:29:18 -05:00
mabashian
0fc0106cc7 Fix dashboard dropdowns after bootstrap upgrade 2019-03-05 11:25:05 -05:00
Daniel Sami
bcdb590a29 404 tests for varied resources 2019-03-05 10:55:39 -05:00
mabashian
6b46c7db8f Fixes linting errors 2019-03-05 08:50:43 -05:00
John Mitchell
8c5bcffd42 add unsaved workflow changes flow 2019-03-04 16:58:10 -05:00
softwarefactory-project-zuul[bot]
aad185e785 Merge pull request #3356 from marshmalien/feat-toolbar-sort-projects-list
Add toolbar sort to projects list

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-04 21:31:44 +00:00
mabashian
2ff22bd681 Show yaml comments when possible on launch prompt 2019-03-04 16:18:23 -05:00
softwarefactory-project-zuul[bot]
2934fabd98 Merge pull request #3357 from mabashian/jobs-list-socket-refresh
Fixes bug where jobs list search was lost on socket message

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-04 20:52:29 +00:00
mabashian
2359231bda Remove console.logs 2019-03-04 15:14:27 -05:00
mabashian
ca5f27aa9e Removes vm.querySet from jobs list and ensures that state params are updated when search is performed to prevent context loss on data refresh. 2019-03-04 15:11:54 -05:00
Marliana Lara
80adcaab81 Set sort dropdown to state param order_by value 2019-03-04 14:40:15 -05:00
Marliana Lara
8100fc1cfb Add toolbar sort configuration to project list 2019-03-04 12:32:25 -05:00
Daniel Sami
006797014c fix url 2019-03-04 11:58:32 -05:00
Daniel Sami
38934dc8d0 lint fixes 2019-03-04 11:56:48 -05:00
Daniel Sami
0ec6d652f7 check for 404s 2019-03-04 11:53:35 -05:00
Marliana Lara
1525c6d97e Extend Toolbar directive to include sort dropdown 2019-03-04 11:25:58 -05:00
softwarefactory-project-zuul[bot]
bf1769af6c Merge pull request #3322 from mopahle/docker_install_ssl_default
Add SSL suport for docker install

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-01 19:54:05 +00:00
softwarefactory-project-zuul[bot]
6384e638f5 Merge pull request #3340 from AlexSCorey/3297-smartInventoryLabel
Add padding to label list tag.

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-03-01 16:20:09 +00:00
Alex Corey
1df5e55a4e add padding to label list tag. 2019-03-01 10:56:08 -05:00
softwarefactory-project-zuul[bot]
d1005f91e7 Merge pull request #3174 from jbradberry/org_hosts_limit
[WIP] Org hosts limit

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-28 21:27:50 +00:00
softwarefactory-project-zuul[bot]
d9451ac12c Merge pull request #3335 from AlexSCorey/3309-mgmtJobTitleStyling
Add margin to Management Jobs Notifications list

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-28 21:12:59 +00:00
mabashian
a82304765d Adds help text to host limit error shown in the output details of inventory sync jobs 2019-02-28 15:54:10 -05:00
mabashian
cc3f2e0819 Properly pass path to error message 2019-02-28 15:54:10 -05:00
mabashian
03c07c0843 Adds error handling to launch api calls 2019-02-28 15:54:10 -05:00
mabashian
1b94b616f0 Default empty max hosts to 0 2019-02-28 15:54:09 -05:00
mabashian
98c5cb1c4c Adds max value to host limit input 2019-02-28 15:54:09 -05:00
mabashian
ce5a85a53b Add support for max hosts on org 2019-02-28 15:54:09 -05:00
Jeff Bradberry
046385d72e Make the inventory_import command log an error message on exception
when it happens in the big try/except block in the middle of handle().
Previously it wasn't doing anything with it, except exiting with a
code of 1.
2019-02-28 15:54:09 -05:00
Jeff Bradberry
7eba55fbde Change the wording of the error when adding a host
to "Organization host limit of %s would be exceeded...", since the
host will probably not actually be made active.
2019-02-28 15:54:09 -05:00
Jeff Bradberry
6ac51b7b13 Update the permission error to include max_hosts and the current host count 2019-02-28 15:54:09 -05:00
Jeff Bradberry
97cc467ae1 Restrict edit permissions on the Organization.max_hosts field to superusers 2019-02-28 15:54:09 -05:00
Jeff Bradberry
3312ebcb05 Remove the hosts count from related_field_counts in the org api endpoints
It is probably not needed, and adds an additional db query.
2019-02-28 15:54:09 -05:00
Jeff Bradberry
4d06ae48d3 Deal with the (erroneous) case where a job is missing the inventory
by bailing out of check_org_host_limit early.  Validation catches this
situation later on.
2019-02-28 15:54:09 -05:00
Jeff Bradberry
cf75ea91a1 Properly use the inventory in the can_start permissions checks 2019-02-28 15:54:09 -05:00
Jeff Bradberry
5e34f6582b Tests for start permissions for JobTemplate and WorkflowJob
when max_hosts is set.
2019-02-28 15:54:09 -05:00
Jeff Bradberry
60008dbd74 Add a test that AWX doesn't care about max_hosts when editing hosts
on an inventory.
2019-02-28 15:54:09 -05:00
Jeff Bradberry
4afd0672a1 Add test that AWX doesn't care about the limit when creating new hosts 2019-02-28 15:54:09 -05:00
Jeff Bradberry
875a1c0b5f Remove the mention of the max_hosts value from the limit check messages 2019-02-28 15:54:09 -05:00
Jeff Bradberry
f5d7ca6913 Update the Organization API list and detail tests to check the host counts 2019-02-28 15:54:09 -05:00
Jeff Bradberry
df8a66e504 Correct the org limit check for changing hosts to use the host's org
instead of an inventory passed in from the user data, which is not allowed.
2019-02-28 15:54:09 -05:00
Jeff Bradberry
43a0a15f6f Expose the new InventoryUpdate.org_host_limit_error field in the API 2019-02-28 15:54:09 -05:00
Jeff Bradberry
36ed890c14 Add permissions checks for the organization host limit 2019-02-28 15:54:09 -05:00
Jeff Bradberry
0e8e5f65e1 Update to fix tests 2019-02-28 15:54:09 -05:00
Jeff Bradberry
6399ec59c9 Include in the API the count of hosts used by an organization 2019-02-28 15:54:09 -05:00
Jeff Bradberry
e44c73883e Expose Organization.max_hosts in the API 2019-02-28 15:54:09 -05:00
Jeff Bradberry
c3406748de Add the new fields to the database 2019-02-28 15:54:03 -05:00
Jeff Bradberry
6d70651611 Update the inventory_import management command
to respect the new Organization.max_hosts limit.
2019-02-28 15:51:50 -05:00
Jeff Bradberry
5e13da62a4 Added a max_hosts field to Organization
in order to optionally set a limit to the number of hosts an
organization that is sharing a license is allowed to manage.

related #1542
2019-02-28 15:51:50 -05:00
softwarefactory-project-zuul[bot]
658e5f0fc8 Merge pull request #3272 from jladdjr/mo_stats
Add support for new playbook stats

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-28 18:24:03 +00:00
softwarefactory-project-zuul[bot]
cebd918e49 Merge pull request #3331 from mabashian/IG-instances-modal
Fixes bug where all instances were preselected when modal was initially opened

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-28 14:30:43 +00:00
Alex Corey
a8c4e92804 Add margin to Management Jobs Notifications list 2019-02-28 09:17:27 -05:00
softwarefactory-project-zuul[bot]
dea71d2682 Merge pull request #3333 from ryanpetrello/more-py3-bugs
fix a few additional py3 issues

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-28 13:31:05 +00:00
Markus Opahle
ed568f569c only use ssl if certificate is specified
Signed-off-by: Markus Opahle <3225748+mopahle@users.noreply.github.com>
2019-02-28 14:06:59 +01:00
Ryan Petrello
13c05c68fc fix a few additional py3 issues
related: https://github.com/ansible/awx/issues/3329
2019-02-27 16:56:36 -05:00
mabashian
2be7f853f3 Fixes bug where all instances were preselected when modal was initially opened 2019-02-27 15:24:29 -05:00
softwarefactory-project-zuul[bot]
3b5681465a Merge pull request #3308 from AlexSCorey/3283-BorderIssues
Removes top left border radius on lists

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-27 16:55:01 +00:00
softwarefactory-project-zuul[bot]
a6c7502217 Merge pull request #3294 from ansible/form-error-handling
Tests for form validation and error checks

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-27 15:58:39 +00:00
softwarefactory-project-zuul[bot]
50a87843ee Merge pull request #3315 from mabashian/elapsed-timer
Show elapsed timer while job is running

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-27 15:54:20 +00:00
Daniel Sami
238b6cbb61 E2E tests for form validation 2019-02-27 10:33:09 -05:00
walkafwalka
3a7bf6a8ac Add SSL suport for docker install
Signed-off-by: walkafwalka <41709139+walkafwalka@users.noreply.github.com>
2019-02-27 10:45:34 +01:00
mabashian
6e64aa81fd Always show elapsed timer regardless of running state 2019-02-26 11:38:46 -05:00
Alex Corey
445612315b Removes top left border radius on list item where the list has a tool bar
Removes redundant styling.
2019-02-26 09:52:07 -05:00
softwarefactory-project-zuul[bot]
bb276a8fcb Merge pull request #3260 from jlmitch5/launchJTWFForm
add launch button to jt and wf forms

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-25 17:52:01 +00:00
Jim Ladd
b7b0bdaeca Ansible 2.8 deprecates use of -U 2019-02-25 00:42:19 -08:00
Jim Ladd
cc1a97b6d8 Update JobHostSummary.__str__ and corresponding tests 2019-02-25 00:42:19 -08:00
Jim Ladd
c6227797b4 Make new host summary fields backwards compatible 2019-02-22 14:07:07 -08:00
Chris Meyers
b383144b69 Merge pull request #3293 from chrismeyersfsu/fix-serial
fix handling of serial strategy
2019-02-22 13:30:24 -05:00
softwarefactory-project-zuul[bot]
d1bc013da9 Merge pull request #3292 from kialam/fix-dummy-data-generator-script-errors
Fix various errors when trying to run `make bulk_data`.

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-22 18:30:06 +00:00
chris meyers
723f581fd0 fix handling of serial strategy
* v2_playbook_on_play_start is called multiple times for the same UUID.
Specifically, once for each host in the play. This changes makes the
uuid unique before going to the dispatcher.
2019-02-22 13:14:35 -05:00
Kia Lam
d2c345a374 Fix API lint error. 2019-02-22 13:06:20 -05:00
softwarefactory-project-zuul[bot]
22677029e1 Merge pull request #3282 from AlexSCorey/2261-LimitHelper
Add helper text to the limit field prompt modal

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-22 17:55:08 +00:00
Kia Lam
0943f989ce Fix various errors when trying to run make bulk_data.
- Properly import PrimoridialModel.
- Use floor division operator for range() method to avoid float vs int errors.
2019-02-22 11:59:18 -05:00
softwarefactory-project-zuul[bot]
c1698fff8e Merge pull request #3264 from kialam/status-service-on-success
Call `this.sync` when we receive a ws complete message.

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-21 21:16:33 +00:00
John Mitchell
200028b269 updated initial tooltip setting 2019-02-21 16:07:27 -05:00
Alex Corey
0cae612159 Add helper text to the limit field prompt modal 2019-02-21 16:06:18 -05:00
softwarefactory-project-zuul[bot]
5584f6a98b Merge pull request #3279 from mabashian/prompt-extra-var-wf
Add prompt for extra vars to wfjt

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-21 18:36:11 +00:00
softwarefactory-project-zuul[bot]
e5acf93c66 Merge pull request #3265 from mabashian/socket-time-band
Prevent jobs/templates lists from refreshing list data more than once every 5 seconds

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-21 18:06:41 +00:00
mabashian
01b4b47087 Fixes unit test failure after adding ui support for extra var prompting on workflows 2019-02-21 12:55:30 -05:00
John Mitchell
14e9923037 remove select2count that wasn't being used 2019-02-21 12:19:45 -05:00
John Mitchell
7e47a924c5 break out credential watch statement in order to disable launch when creds are removed 2019-02-21 12:18:34 -05:00
John Mitchell
635aa9fd56 promise-ify createselect2 and use that instead for deferring launch button enabling 2019-02-21 12:18:34 -05:00
John Mitchell
2e4eb1885f add launch button to jt and wf forms 2019-02-21 12:18:33 -05:00
mabashian
0ce70c08bd Add prompt for extra vars to wfjt 2019-02-21 11:37:53 -05:00
mabashian
e3c9a42741 Hook up status update sockets on completed jobs routes and project templates 2019-02-21 11:24:37 -05:00
mabashian
d5d45e644d Prevent jobs/templates lists from refreshing list data more than once every 5 seconds 2019-02-21 11:24:37 -05:00
softwarefactory-project-zuul[bot]
09684e2c41 Merge pull request #3269 from ryanpetrello/dep-update
remove redbaron and update minor dependencies

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-21 15:56:02 +00:00
Ryan Petrello
04622d5786 remove redbaron and update dependencies 2019-02-21 10:08:24 -05:00
Jim Ladd
8c9544e5ed Add support for new ansible stats 2019-02-20 17:13:29 -08:00
softwarefactory-project-zuul[bot]
ca043d9bfd Merge pull request #3275 from AlanCoding/i_love_caches
Clear the test cache at end of every test

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-20 20:19:16 +00:00
AlanCoding
711937b104 fix some patches that were never unapplied 2019-02-20 14:40:25 -05:00
softwarefactory-project-zuul[bot]
a9a2c1fa7b Merge pull request #3245 from wenottingham/we-meet-again-for-the-last-login-time
[WIP] Show last login along with the badges on the user view screen.

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-20 19:38:07 +00:00
kialam
ec8a452f1d Call this.sync when we receive a ws complete message.
- There can be a race condition where a job run finishes event processing before sending off a "successful/failed" websocket message. In this instance, we would miss gathering known artifacts for a job. The proposed fix is to do a manual detection of when a job status updates to "successful/failed/unknown" and perform a GET request to the job endpoint.
2019-02-20 14:11:13 -05:00
AlanCoding
07def62373 clear the test cache at end of every test 2019-02-20 14:02:36 -05:00
softwarefactory-project-zuul[bot]
30352e375f Merge pull request #3271 from kialam/compact-expand-lists-responsive
Fix Compact/Expand Lists responsive issues

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-20 16:59:40 +00:00
softwarefactory-project-zuul[bot]
6abe9d5c0f Merge pull request #3270 from saito-hideki/issue/tower_3344
Add action to output login failure event to logfile

Reviewed-by: Ryan Petrello
             https://github.com/ryanpetrello
2019-02-20 16:01:31 +00:00
softwarefactory-project-zuul[bot]
9b04e93765 Merge pull request #3262 from ryanpetrello/openstack_cred_verify_ssl
add verify_ssl to OpenStack credential type

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-20 15:35:10 +00:00
kialam
728fb1aaef Fix hidden relaunch dropdown in new lists. 2019-02-20 10:05:21 -05:00
kialam
2596ad26b9 Make compact rows responsive. 2019-02-20 09:35:22 -05:00
Hideki Saito
ef3b1ee195 Add action to output login failure to logger
Signed-off-by: Hideki Saito <saito@fgrep.org>
2019-02-20 14:27:44 +00:00
Ryan Petrello
b1a33869dc convey OpenStack verify_ssl defaults in the CredentialType schema 2019-02-20 09:02:48 -05:00
Hideki Saito
9f04fbe4a4 Add verify_ssl to OpenStack credential type
To avoid verification failures when using a self-signed certificate file,
 Added "Verify SSL" check box to the openstack credential type edit page.

Signed-off-by: Hideki Saito <saito@fgrep.org>
2019-02-19 12:53:13 -05:00
softwarefactory-project-zuul[bot]
1ece764547 Merge pull request #3257 from ryanpetrello/native_credential_types
define native CredentialType inputs/injectors in code, not in the DB

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-19 17:27:19 +00:00
softwarefactory-project-zuul[bot]
2358e306c1 Merge pull request #3261 from elyezer/resize-window
Resize window to avoid breaking the UI

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-19 17:21:12 +00:00
Elyézer Rezende
82b9f8ebb0 Resize window to avoid breaking the UI
Update test-jobs-portal-list-actions to resize the browser window and
avoid a flake test because of broken layout.
2019-02-19 13:53:12 -03:00
Ryan Petrello
43ca4526b1 define native CredentialType inputs/injectors in code, not in the DB
This has a few benefits:

1.  It makes adding new fields to built-in CredentialTypes _much_
    simpler.  In the past, we've had to write a migration every time we
    want to modify an existing type (changing a label/help text,
    changing options like the recent become_method changes) or
    when adding a new field entirely

2.  It paves the way for third party credential plugins support, where
    importable libraries will define their own source code-based schema
2019-02-19 10:22:26 -05:00
softwarefactory-project-zuul[bot]
4174fc22b0 Merge pull request #3254 from vismay-golwala/job_host_count
Fixed # of hosts - job details UI

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-18 21:16:32 +00:00
Vismay Golwala
16e135249c Fixed # of hosts - job details UI
In the UI job details page, the number of hosts for any job was always displayed '1'
irrespective of the actual count. This was caused because of a faulty initialization
of variable followed by unreachable code. It has been fixed by updating init value.

Signed-off-by: Vismay Golwala <vgolwala@redhat.com>
2019-02-18 13:57:58 -05:00
softwarefactory-project-zuul[bot]
889dae357b Merge pull request #3235 from ryanpetrello/sql-profiling
add a custom DB backend that provides system-level SQL profiling

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-15 21:56:28 +00:00
softwarefactory-project-zuul[bot]
0063668582 Merge pull request #3247 from ryanpetrello/fix-serial-display-error
fix a bug in the display of playbooks using serial or free strategy

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-15 21:51:54 +00:00
Ryan Petrello
1e4cd9ea8f document the SQL profiler 2019-02-15 16:34:34 -05:00
Ryan Petrello
954ccccbc5 fix a bug in the display of playbooks using serial or free strategy 2019-02-15 16:13:12 -05:00
Ryan Petrello
6f43875e80 record profile data in /var/log/tower, not /var/lib/awx 2019-02-15 14:34:55 -05:00
Christian Adams
80cccab919 Merge pull request #3244 from e-tienne/fix_wording_wf
Fix ambiguous workflow vizualizer 5000 wording
2019-02-15 11:23:58 -05:00
softwarefactory-project-zuul[bot]
088673ceb0 Merge pull request #3246 from Klaas-/Klaas--patch-2
Avoid pg password ending up in syslog/shell output

Reviewed-by: awxbot
             https://github.com/awxbot
2019-02-15 15:48:39 +00:00
Vismay Golwala
4f13255f35 Allow empty default values for numerical survey answers.
Signed-off-by: Vismay Golwala <vgolwala@redhat.com>
2019-02-15 10:41:19 -05:00
Klaas Demter
8f36e21c97 Avoid pg password ending up in syslog/shell output
Currently if an error occurs the pgpassword would be exposed to syslog / shell during playbook backup.yml
2019-02-15 16:15:33 +01:00
Bill Nottingham
6a11281355 Show last login along with the badges on the user view screen. 2019-02-14 20:25:50 -05:00
Etienne Simard
df4e4f80ad Fix ambiguous workflow vizualizer wording
Resolves: #2998

Signed-off-by: Etienne Simard <etienne@redhat.com>
2019-02-14 18:01:23 -05:00
softwarefactory-project-zuul[bot]
5682fb1503 Merge pull request #3243 from mabashian/more-strings-to-translate
Translate job status in smart status tooltip.  Mark strings for translation in project form

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-14 22:46:25 +00:00
mabashian
640d2a2797 Fix linting error 2019-02-14 16:32:54 -05:00
softwarefactory-project-zuul[bot]
b173880766 Merge pull request #3240 from wenottingham/when-last-we-met
Add django last_login information to user object.

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-14 21:32:09 +00:00
mabashian
30ce85b80a Translate job status in smart status tooltip. Mark strings for translation in project form 2019-02-14 16:20:33 -05:00
softwarefactory-project-zuul[bot]
003ec64413 Merge pull request #3241 from ryanpetrello/workflow-convergence-i18n
mark a workflow convergence error message for i18n

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-14 21:16:54 +00:00
softwarefactory-project-zuul[bot]
eda6d729d6 Merge pull request #3239 from wenottingham/did-this-ever-work
Remove `awx-manage user_info` command.

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-14 21:12:50 +00:00
Ryan Petrello
4f83d44142 mark a workflow convergence error message for i18n 2019-02-14 15:55:20 -05:00
Bill Nottingham
7452eb2fa1 Remove broken user_info command.
This has not worked in a long time, and does not serve much purpose.
2019-02-14 15:34:24 -05:00
Bill Nottingham
8300f7f51b Add django last_login information to user object. 2019-02-14 15:17:37 -05:00
Ryan Petrello
eed94b641e add a custom DB backend that provides system-level SQL profiling
run this command on _any_ node in an awx cluster:

$ awx-manage profile_sql --threshold=2.0 --minutes=1

...and for 1 minute, the timing for _every_ SQL query in _every_ awx
Python process that uses the Django ORM will be measured

queries that run longer than (in this example) 2 seconds will be
written to a per-process sqlite database in /var/lib/awx/profile, and
the file will contain an EXPLAIN VERBOSE for the query and the full
Python stack that led to that SQL query's execution (this includes not
just WSGI requests, but background processes like the runworker and
dispatcher)

$ awx-manage profile_sql --threshold=0

...can be used to disable profiling again (if you don't want to wait for
the minute to expire)
2019-02-14 15:04:46 -05:00
Unknown
0138e92ddc update documentation to include kuberentes initContainers
Update documentation to include Kubernetes initContainers in custom virtualenvs
2019-02-14 14:07:26 -05:00
softwarefactory-project-zuul[bot]
456ef49ee3 Merge pull request #3229 from mabashian/popover-quote-entity
Replace single quote with appropriate entity when generating new attribute

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-14 16:32:57 +00:00
softwarefactory-project-zuul[bot]
b91dee68ac Merge pull request #3221 from mabashian/workflow-results-inv-tooltip
Show the proper tooltip on workflow results inventory

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-14 16:28:34 +00:00
softwarefactory-project-zuul[bot]
781d36ef83 Merge pull request #3220 from mabashian/credential-modal-400
Prevent extra fetch of cred list as cred modal is closing

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-14 16:06:52 +00:00
softwarefactory-project-zuul[bot]
a1cef744a7 Merge pull request #3230 from impca/patch-1
Update compose configuration

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-14 15:45:41 +00:00
softwarefactory-project-zuul[bot]
ba5319f479 Merge pull request #3213 from mabashian/3158-related-groups-style
Fix styling on related groups labels after bootstrap upgrade

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-14 15:40:56 +00:00
softwarefactory-project-zuul[bot]
0dbf21a15c Merge pull request #3176 from digipok/issue-3010-ca-trust-awx_task
update-ca-trust: Ensure CA trust is updated in awx_task container

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-14 15:36:55 +00:00
softwarefactory-project-zuul[bot]
45d522829a Merge pull request #3190 from mabashian/i18n-strings
Mark various strings for translation

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-14 15:26:50 +00:00
softwarefactory-project-zuul[bot]
8b1c358dc6 Merge pull request #3165 from mabashian/2630-become-plugins
Makes priv escalation method a dynamic select element

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-14 15:08:09 +00:00
softwarefactory-project-zuul[bot]
ebd9d3dc67 Merge pull request #3234 from wenottingham/the-only-good-code-is-a-dead-code
Delete some unused functions.

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-14 14:10:31 +00:00
softwarefactory-project-zuul[bot]
80cf154fb7 Merge pull request #3233 from ryanpetrello/F405
remove usage of import * and enforce F405 in our linter

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-14 13:32:18 +00:00
impca
9add96a0d3 update docker compose installer
Only run commands to update certs when config changes.
2019-02-14 08:29:47 +01:00
mickfeech
ed2ad1e210 update documentation to include kuberentes initContainers 2019-02-13 20:01:30 -05:00
softwarefactory-project-zuul[bot]
808ed74700 Merge pull request #3177 from wenottingham/i-am-going-to-need-a-diagram-for-this
Fix project updates to properly pull in role requirements.

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-13 23:16:34 +00:00
Ryan Petrello
9bebf3217e remove usage of import * and enforce F405 in our linter
import * is a scourge upon the earth
2019-02-13 17:10:33 -05:00
softwarefactory-project-zuul[bot]
ae7d26fab0 Merge pull request #3186 from vismay-golwala/update_schedule_constraint
Update standalone schedule name uniqueness combining it with unified …

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-13 22:03:54 +00:00
Bill Nottingham
0f54d30f2c Remove some unused functions. 2019-02-13 16:29:59 -05:00
softwarefactory-project-zuul[bot]
631d3515f2 Merge pull request #3217 from mabashian/3164-cred-checkbox-alignment
Fixes cred form checkbox input styling

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-13 20:08:57 +00:00
softwarefactory-project-zuul[bot]
551218fd44 Merge pull request #3206 from AlanCoding/learn_to_share
Do not remove edges from other inventory sources

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-13 18:09:18 +00:00
Vismay Golwala
4af54517d2 Update standalone schedule name uniqueness combining it with unified job template.
Signed-off-by: Vismay Golwala <vgolwala@redhat.com>
2019-02-13 12:46:28 -05:00
mabashian
334f571ad3 Fixes bug where extra blank option was being added to select input 2019-02-13 11:48:43 -05:00
softwarefactory-project-zuul[bot]
295afa805c Merge pull request #3212 from AlanCoding/model_star_imports
Remove star imports in tasks and non-base models

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-13 15:05:01 +00:00
softwarefactory-project-zuul[bot]
ad4d286db5 Merge pull request #3225 from ryanpetrello/take-it-to-the-limit-one-more-time
remove field size limit on adhoc `limit`

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-13 14:50:56 +00:00
softwarefactory-project-zuul[bot]
eea97c8928 Merge pull request #3228 from jakemcdermott/dev-migrations-page
load migrations page in dev environment

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-13 14:44:53 +00:00
impca
c29275315e Update compose configuration
When running awx via docker-compose and using custom certificates (for LDAP auth or whatever else...), update-ca-trust has to be called afer starting the container to actually use new certificates (just as it is called when using docker to run - https://github.com/ansible/awx/blob/devel/installer/roles/local_docker/tasks/standalone.yml#L119-L120 ).
2019-02-13 15:39:52 +01:00
mabashian
ef89195e6c Replace single quote with appropriate entity when generating new attribute 2019-02-13 09:32:46 -05:00
softwarefactory-project-zuul[bot]
06ff26752a Merge pull request #3227 from ryanpetrello/fix-migration-tran-view
fix `/migrations_notran`

Reviewed-by: awxbot
             https://github.com/awxbot
2019-02-13 14:25:29 +00:00
softwarefactory-project-zuul[bot]
58f5e1882e Merge pull request #3218 from kialam/fix-3081-copy-project-expand-collapse
Fix expanded view not persisting when a user copies a project.

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-13 14:24:50 +00:00
Jake McDermott
2765367308 load migrations page in dev environment 2019-02-13 09:14:11 -05:00
Ryan Petrello
2a94611801 fix /migrations_notran 2019-02-13 08:59:20 -05:00
Ryan Petrello
e4eda3ef0d remove field size limit on adhoc limit
related: 4b669fb16d
2019-02-13 08:34:10 -05:00
AlanCoding
fbf6315a8c remove star imports in tasks and non-base models 2019-02-12 19:50:30 -05:00
softwarefactory-project-zuul[bot]
8a3c10686e Merge pull request #3191 from chrismeyersfsu/fix-job_event_smart_inv_slow_take_two_devel
do not observe queries when constructing them

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-12 22:27:11 +00:00
chris meyers
c121565209 add tests for host_filter
* Ensure that building the smart inventory query string doesn't invoke
the database.
2019-02-12 16:11:54 -05:00
mabashian
3c3e659042 Show the proper tooltip on workflow results inventory 2019-02-12 15:50:52 -05:00
mabashian
406cb07018 Prevent extra fetch of cred list as cred modal is closing 2019-02-12 15:37:45 -05:00
kialam
099a82fdf8 Fix expanded view not persisting when a user copies a project. 2019-02-12 15:24:17 -05:00
softwarefactory-project-zuul[bot]
52b88d839e Merge pull request #3216 from ryanpetrello/inv-gen-speedup
optimize a slow query in inventory script generation

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-12 18:48:30 +00:00
Ryan Petrello
e245e50ee4 optimize a slow query in inventory script generation
if we don't preload this column, Django needs it, and so it generates
one query per-host (!!!) to get it.  For large (10k+ host) inventories,
this is incredibly slow.

see: https://github.com/ansible/awx/issues/3214
2019-02-12 12:55:38 -05:00
mabashian
a52c0415d9 Copied old bootstrap styles to display cred form checkboxes as they were before bootstrap upgrade 2019-02-12 11:48:48 -05:00
mabashian
98c7df3399 Fix styling on related groups labels after bootstrap upgrade 2019-02-12 11:18:28 -05:00
softwarefactory-project-zuul[bot]
570283fba2 Merge pull request #3207 from mabashian/pull-strings-ejs
Extract translations from ejs files

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-11 18:29:55 +00:00
AlanCoding
1bf2a455c6 Do not remove edges from other inventory sources 2019-02-11 13:08:19 -05:00
mabashian
cb222aaa40 Extract translations from ejs files 2019-02-11 12:59:47 -05:00
softwarefactory-project-zuul[bot]
6c9fc4a592 Merge pull request #3198 from mabashian/workflow-link-tooltip-overflow
Dynamically place link hover tooltip based on its size

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-11 17:01:08 +00:00
softwarefactory-project-zuul[bot]
53a6341320 Merge pull request #3148 from elyezer/e2e-screenshots
Add settings to control screenshot capturing

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-11 16:47:06 +00:00
mabashian
c0f9ee5e6e Fixes linting errors 2019-02-11 11:38:28 -05:00
mabashian
3321f3e34d Fixes bug where tooltip was clipped when graph is zoomed out. Fixes linting error 2019-02-11 11:30:20 -05:00
Elyézer Rezende
953b6679ef Add settings to control screenshot capturing 2019-02-11 14:05:11 -02:00
mabashian
d285261697 Fixes split job unit test confirming string match 2019-02-11 10:28:35 -05:00
softwarefactory-project-zuul[bot]
d84b58c857 Merge pull request #3196 from kialam/detect-empty-artifacts
Job Artifacts: check for empty object rather than string value.

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-11 15:21:07 +00:00
softwarefactory-project-zuul[bot]
4d3cacf87f Merge pull request #3185 from mabashian/edit-node-arrow
Use arrow instead of 'to' in workflow edit link title

Reviewed-by: https://github.com/softwarefactory-project-zuul[bot]
2019-02-11 14:51:05 +00:00
kialam
fcba02cd86 Linter fix. 2019-02-08 14:52:15 -07:00
mabashian
205dc93e65 Dynamically place link hover tooltip based on its size 2019-02-08 16:44:26 -05:00
kialam
28a29293c7 Check for empty object rather than string value. 2019-02-08 14:31:06 -07:00
chris meyers
ed78978b5f do not observe queries when constructing them
* While parsing host_filter in the smart inventory code it was
triggering sql queries. This changset avoids executing the query that is
being constructed.
2019-02-08 12:30:51 -05:00
mabashian
6e1457607e Mark various strings for translation 2019-02-08 11:28:04 -05:00
mabashian
b4d54895ff Use arrow instead of 'to' in workflow edit link title 2019-02-07 13:46:22 -05:00
Bill Nottingham
5df384edd6 Fix project updates to properly pull in role requirements.
"check" runs check out the version that is saved in the database,
so for git repos, any subsequent "checkout" run on the same node
would always report that we have the proper version, and we would
not properly force a role update when requiremets.yml changes.

Also, don't use `scm_result` for all SCMs, as the skipped tasks will
overwrite earlier `scm_result` variables.
2019-02-06 13:20:09 -05:00
Mathieu Mallet
dce3795e0c update-ca-trust: Ensure CA trust is updated in awx_task container
Related #3010

Both awx_web and awx_task containers can have a volume mounted in
specified by the ca_trust_dir variable. Unfortunately only the
awx_web container's trust is updated. This patch makes sure the
awx_task container's trust is updated as well

Testing Done: ansible-playbook --syntax-check installer/install.yml

Signed-off-by: Mathieu Mallet <mmallet@digipok.io>
2019-02-06 16:51:14 +00:00
mabashian
1a619de91f Makes priv escalation method a dynamic select element 2019-02-05 14:01:29 -05:00
749 changed files with 22449 additions and 18179 deletions

View File

@@ -1,6 +0,0 @@
[mini_dinstall]
fqdn = localhost
method = local
incoming = FIXME/deb-repo/mini-dinstall/incoming
run_dinstall = 0
post_upload_command = mini-dinstall -b -v

View File

@@ -1,20 +0,0 @@
root = true
[*]
end_of_line = lf
insert_final_newline = true
[Makefile]
indent_style = tab
[**.py]
indent_style = space
indent_size = 4
[**.{js,less,html}]
indent_style = space
indent_size = 4
[**.{json}]
indent_style = space
indent_size = 2

1
.env
View File

@@ -1,2 +1,3 @@
PYTHONUNBUFFERED=true
SELENIUM_DOCKER_TAG=latest

View File

@@ -7,12 +7,6 @@ about: Create a report to help us improve
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
<!-- Pick the area of AWX for this issue, you can have multiple, delete the rest: -->
- API
- UI
- Installer
##### SUMMARY
<!-- Briefly describe the problem. -->

View File

@@ -7,16 +7,5 @@ about: Suggest an idea for this project
##### ISSUE TYPE
- Feature Idea
##### COMPONENT NAME
<!-- Pick the area of AWX for this issue, you can have multiple, delete the rest: -->
- API
- UI
- Installer
##### SUMMARY
<!-- Briefly describe the problem or desired enhancement. -->
##### ADDITIONAL INFORMATION
<!-- Include any links to sosreport, database dumps, screenshots or other
information. -->

2
.gitignore vendored
View File

@@ -8,6 +8,7 @@ reference-schema.json
.tags1
# Tower
awx-dev
awx/settings/local_*.py*
awx/*.sqlite3
awx/*.sqlite3_*
@@ -29,6 +30,7 @@ awx/ui/templates/ui/index.html
awx/ui/templates/ui/installing.html
/tower-license
/tower-license/**
tools/prometheus/data
# Tower setup playbook testing
setup/test/roles/postgresql

View File

@@ -1,16 +0,0 @@
[DEFAULT]
archivedir = FIXME/deb-repo
mail_to =
verify_sigs = false
architectures = all, amd64
archive_style = flat
generate_release = true
mail_on_success = false
release_codename = ansible-tower
release_description = Ansible Tower
release_label = ansible-tower
release_origin = ansible-tower
[trusty]
[precise]

View File

@@ -1,6 +0,0 @@
[MASTER]
# Add files or directories to the blacklist. They should be base names, not
# paths.
ignore=site-packages,ui,migrations,data

View File

@@ -83,12 +83,10 @@ If you're not using Docker for Mac, or Docker for Windows, you may need, or choo
(host)$ pip install docker-compose
```
#### Node and npm
#### Frontend Development
The AWX UI requires the following:
See [the ui development documentation](awx/ui/README.md).
- Node 8.x LTS
- NPM 6.x LTS
### Build the environment

View File

@@ -27,7 +27,7 @@ This document provides a guide for installing AWX.
- [Start the build](#start-the-build-1)
- [Accessing AWX](#accessing-awx-1)
- [SSL Termination](#ssl-termination)
- [Docker or Docker Compose](#docker-or-docker-compose)
- [Docker Compose](#docker-compose)
- [Prerequisites](#prerequisites-3)
- [Pre-build steps](#pre-build-steps-2)
- [Deploying to a remote host](#deploying-to-a-remote-host)
@@ -59,10 +59,13 @@ Before you can run a deployment, you'll need the following installed in your loc
- [Ansible](http://docs.ansible.com/ansible/latest/intro_installation.html) Requires Version 2.4+
- [Docker](https://docs.docker.com/engine/installation/)
- [docker-py](https://github.com/docker/docker-py) Python module
+ A recent version
- [docker](https://pypi.org/project/docker/) Python module
+ This is incompatible with `docker-py`. If you have previously installed `docker-py`, please uninstall it.
+ We use this module instead of `docker-py` because it is what the `docker-compose` Python module requires.
- [GNU Make](https://www.gnu.org/software/make/)
- [Git](https://git-scm.com/) Requires Version 1.8.4+
- [Node 8.x LTS version](https://nodejs.org/en/download/)
- [Node 10.x LTS version](https://nodejs.org/en/download/)
- [NPM 6.x LTS](https://docs.npmjs.com/)
### System Requirements
@@ -73,7 +76,7 @@ The system that runs the AWX service will need to satisfy the following requirem
- At least 2 cpu cores
- At least 20GB of space
- Running Docker, Openshift, or Kubernetes
- If you choose to use an external PostgreSQL database, please note that the minimum version is 9.4.
- If you choose to use an external PostgreSQL database, please note that the minimum version is 9.6+.
### AWX Tunables
@@ -81,14 +84,14 @@ The system that runs the AWX service will need to satisfy the following requirem
### Choose a deployment platform
We currently support running AWX as a containerized application using Docker images deployed to either an OpenShift cluster, docker-compose or a standalone Docker daemon. The remainder of this document will walk you through the process of building the images, and deploying them to either platform.
We currently support running AWX as a containerized application using Docker images deployed to either an OpenShift cluster, a Kubernetes cluster, or docker-compose. The remainder of this document will walk you through the process of building the images, and deploying them to either platform.
The [installer](./installer) directory contains an [inventory](./installer/inventory) file, and a playbook, [install.yml](./installer/install.yml). You'll begin by setting variables in the inventory file according to the platform you wish to use, and then you'll start the image build and deployment process by running the playbook.
In the sections below, you'll find deployment details and instructions for each platform:
- [OpenShift](#openshift)
- [Kubernetes](#kubernetes)
- [Docker or Docker Compose](#docker-or-docker-compose).
- [Docker Compose](#docker-compose).
### Official vs Building Images
@@ -391,14 +394,14 @@ If your provider is able to allocate an IP Address from the Ingress controller t
Unlike Openshift's `Route` the Kubernetes `Ingress` doesn't yet handle SSL termination. As such the default configuration will only expose AWX through HTTP on port 80. You are responsible for configuring SSL support until support is added (either to Kubernetes or AWX itself).
## Docker or Docker-Compose
## Docker-Compose
### Prerequisites
- [Docker](https://docs.docker.com/engine/installation/) on the host where AWX will be deployed. After installing Docker, the Docker service must be started (depending on your OS, you may have to add the local user that uses Docker to the ``docker`` group, refer to the documentation for details)
- [docker-py](https://github.com/docker/docker-py) Python module.
If you're installing using Docker Compose, you'll need [Docker Compose](https://docs.docker.com/compose/install/).
- [docker-compose](https://pypi.org/project/docker-compose/) Python module.
+ This also installs the `docker` Python module, which is incompatible with `docker-py`. If you have previously installed `docker-py`, please uninstall it.
- [Docker Compose](https://docs.docker.com/compose/install/).
### Pre-build steps
@@ -441,13 +444,17 @@ Before starting the build process, review the [inventory](./installer/inventory)
> Provide a port number that can be mapped from the Docker daemon host to the web server running inside the AWX container. Defaults to *80*.
*use_docker_compose*
*host_port_ssl*
> Switch to ``true`` to use Docker Compose instead of the standalone Docker install.
> Provide a port number that can be mapped from the Docker daemon host to the web server running inside the AWX container for SSL support. Defaults to *443*, only works if you also set `ssl_certificate` (see below).
*ssl_certificate*
> Optionally, provide the path to a file that contains a certificate and its private key.
*docker_compose_dir*
When using docker-compose, the `docker-compose.yml` file will be created there (default `/var/lib/awx`).
> When using docker-compose, the `docker-compose.yml` file will be created there (default `/tmp/awxcompose`).
*ca_trust_dir*
@@ -527,7 +534,7 @@ After the playbook run completes, Docker will report up to 5 running containers.
```bash
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
e240ed8209cd awx_task:1.0.0.8 "/tini -- /bin/sh ..." 2 minutes ago Up About a minute 8052/tcp awx_task
1cfd02601690 awx_web:1.0.0.8 "/tini -- /bin/sh ..." 2 minutes ago Up About a minute 0.0.0.0:80->8052/tcp awx_web
1cfd02601690 awx_web:1.0.0.8 "/tini -- /bin/sh ..." 2 minutes ago Up About a minute 0.0.0.0:443->8052/tcp awx_web
55a552142bcd memcached:alpine "docker-entrypoint..." 2 minutes ago Up 2 minutes 11211/tcp memcached
84011c072aad rabbitmq:3 "docker-entrypoint..." 2 minutes ago Up 2 minutes 4369/tcp, 5671-5672/tcp, 25672/tcp rabbitmq
97e196120ab3 postgres:9.6 "docker-entrypoint..." 2 minutes ago Up 2 minutes 5432/tcp postgres

View File

@@ -60,7 +60,7 @@ I18N_FLAG_FILE = .i18n_built
.PHONY: awx-link clean clean-tmp clean-venv requirements requirements_dev \
develop refresh adduser migrate dbchange dbshell runserver \
receiver test test_unit test_ansible test_coverage coverage_html \
receiver test test_unit test_coverage coverage_html \
dev_build release_build release_clean sdist \
ui-docker-machine ui-docker ui-release ui-devel \
ui-test ui-deps ui-test-ci VERSION
@@ -124,8 +124,8 @@ virtualenv_ansible:
if [ ! -d "$(VENV_BASE)/ansible" ]; then \
virtualenv -p python --system-site-packages $(VENV_BASE)/ansible && \
$(VENV_BASE)/ansible/bin/pip install $(PIP_OPTIONS) --ignore-installed six packaging appdirs && \
$(VENV_BASE)/ansible/bin/pip install $(PIP_OPTIONS) --ignore-installed setuptools==36.0.1 && \
$(VENV_BASE)/ansible/bin/pip install $(PIP_OPTIONS) --ignore-installed pip==9.0.1; \
$(VENV_BASE)/ansible/bin/pip install $(PIP_OPTIONS) --ignore-installed setuptools==41.0.1 && \
$(VENV_BASE)/ansible/bin/pip install $(PIP_OPTIONS) --ignore-installed pip==19.1.1; \
fi; \
fi
@@ -134,8 +134,8 @@ virtualenv_ansible_py3:
if [ ! -d "$(VENV_BASE)" ]; then \
mkdir $(VENV_BASE); \
fi; \
if [ ! -d "$(VENV_BASE)/ansible3" ]; then \
python3 -m venv --system-site-packages $(VENV_BASE)/ansible3; \
if [ ! -d "$(VENV_BASE)/ansible" ]; then \
$(PYTHON) -m venv --system-site-packages $(VENV_BASE)/ansible; \
fi; \
fi
@@ -145,7 +145,8 @@ virtualenv_awx:
mkdir $(VENV_BASE); \
fi; \
if [ ! -d "$(VENV_BASE)/awx" ]; then \
$(PYTHON) -m venv $(VENV_BASE)/awx; \
$(PYTHON) -m venv --system-site-packages $(VENV_BASE)/awx; \
$(VENV_BASE)/awx/bin/pip install $(PIP_OPTIONS) --ignore-installed docutils==0.14; \
fi; \
fi
@@ -158,22 +159,18 @@ requirements_ansible: virtualenv_ansible
$(VENV_BASE)/ansible/bin/pip uninstall --yes -r requirements/requirements_ansible_uninstall.txt
requirements_ansible_py3: virtualenv_ansible_py3
cat requirements/requirements_ansible.txt requirements/requirements_ansible_git.txt | $(VENV_BASE)/ansible3/bin/pip3 install $(PIP_OPTIONS) --no-binary $(SRC_ONLY_PKGS) --ignore-installed -r /dev/stdin
$(VENV_BASE)/ansible3/bin/pip3 install ansible # can't inherit from system ansible, it's py2
$(VENV_BASE)/ansible3/bin/pip3 uninstall --yes -r requirements/requirements_ansible_uninstall.txt
if [[ "$(PIP_OPTIONS)" == *"--no-index"* ]]; then \
cat requirements/requirements_ansible.txt requirements/requirements_ansible_local.txt | $(VENV_BASE)/ansible/bin/pip3 install $(PIP_OPTIONS) --ignore-installed -r /dev/stdin ; \
else \
cat requirements/requirements_ansible.txt requirements/requirements_ansible_git.txt | $(VENV_BASE)/ansible/bin/pip3 install $(PIP_OPTIONS) --no-binary $(SRC_ONLY_PKGS) --ignore-installed -r /dev/stdin ; \
fi
$(VENV_BASE)/ansible/bin/pip3 uninstall --yes -r requirements/requirements_ansible_uninstall.txt
requirements_ansible_dev:
if [ "$(VENV_BASE)" ]; then \
$(VENV_BASE)/ansible/bin/pip install pytest mock; \
fi
requirements_isolated:
if [ ! -d "$(VENV_BASE)/awx" ]; then \
$(PYTHON) -m venv $(VENV_BASE)/awx; \
fi;
echo "include-system-site-packages = true" >> $(VENV_BASE)/awx/lib/python$(PYTHON_VERSION)/pyvenv.cfg
$(VENV_BASE)/awx/bin/pip install -r requirements/requirements_isolated.txt
# Install third-party requirements needed for AWX's environment.
requirements_awx: virtualenv_awx
if [[ "$(PIP_OPTIONS)" == *"--no-index"* ]]; then \
@@ -189,7 +186,7 @@ requirements_awx_dev:
requirements: requirements_ansible requirements_awx
requirements_dev: requirements requirements_ansible_py3 requirements_awx_dev requirements_ansible_dev
requirements_dev: requirements requirements_awx_dev requirements_ansible_dev
requirements_test: requirements
@@ -221,7 +218,7 @@ init:
if [ "$(AWX_GROUP_QUEUES)" == "tower,thepentagon" ]; then \
$(MANAGEMENT_COMMAND) provision_instance --hostname=isolated; \
$(MANAGEMENT_COMMAND) register_queue --queuename='thepentagon' --hostnames=isolated --controller=tower; \
$(MANAGEMENT_COMMAND) generate_isolated_key > /awx_devel/awx/main/expect/authorized_keys; \
$(MANAGEMENT_COMMAND) generate_isolated_key > /awx_devel/awx/main/isolated/authorized_keys; \
fi;
# Refresh development environment after pulling new code.
@@ -272,15 +269,7 @@ supervisor:
@if [ "$(VENV_BASE)" ]; then \
. $(VENV_BASE)/awx/bin/activate; \
fi; \
supervisord --pidfile=/tmp/supervisor_pid
# Alternate approach to tmux to run all development tasks specified in
# Procfile.
honcho:
@if [ "$(VENV_BASE)" ]; then \
. $(VENV_BASE)/awx/bin/activate; \
fi; \
honcho start -f tools/docker-compose/Procfile
supervisord --pidfile=/tmp/supervisor_pid -n
collectstatic:
@if [ "$(VENV_BASE)" ]; then \
@@ -292,7 +281,7 @@ uwsgi: collectstatic
@if [ "$(VENV_BASE)" ]; then \
. $(VENV_BASE)/awx/bin/activate; \
fi; \
uwsgi -b 32768 --socket 127.0.0.1:8050 --module=awx.wsgi:application --home=/venv/awx --chdir=/awx_devel/ --vacuum --processes=5 --harakiri=120 --master --no-orphans --py-autoreload 1 --max-requests=1000 --stats /tmp/stats.socket --lazy-apps --logformat "%(addr) %(method) %(uri) - %(proto) %(status)" --hook-accepting1-once="exec:awx-manage run_dispatcher --reload"
uwsgi -b 32768 --socket 127.0.0.1:8050 --module=awx.wsgi:application --home=/venv/awx --chdir=/awx_devel/ --vacuum --processes=5 --harakiri=120 --master --no-orphans --py-autoreload 1 --max-requests=1000 --stats /tmp/stats.socket --lazy-apps --logformat "%(addr) %(method) %(uri) - %(proto) %(status)" --hook-accepting1="exec:supervisorctl restart tower-processes:awx-dispatcher tower-processes:awx-receiver"
daphne:
@if [ "$(VENV_BASE)" ]; then \
@@ -356,7 +345,8 @@ pylint: reports
@(set -o pipefail && $@ | reports/$@.report)
genschema: reports
$(MAKE) swagger PYTEST_ARGS="--genschema"
$(MAKE) swagger PYTEST_ARGS="--genschema --create-db "
mv swagger.json schema.json
swagger: reports
@if [ "$(VENV_BASE)" ]; then \
@@ -381,20 +371,12 @@ test:
PYTHONDONTWRITEBYTECODE=1 py.test -p no:cacheprovider -n auto $(TEST_DIRS)
awx-manage check_migrations --dry-run --check -n 'vNNN_missing_migration_file'
test_combined: test_ansible test
test_unit:
@if [ "$(VENV_BASE)" ]; then \
. $(VENV_BASE)/awx/bin/activate; \
fi; \
py.test awx/main/tests/unit awx/conf/tests/unit awx/sso/tests/unit
test_ansible:
@if [ "$(VENV_BASE)" ]; then \
. $(VENV_BASE)/ansible/bin/activate; \
fi; \
py.test awx/lib/tests -c awx/lib/tests/pytest.ini
# Run all API unit tests with coverage enabled.
test_coverage:
@if [ "$(VENV_BASE)" ]; then \
@@ -515,6 +497,10 @@ ui-devel: $(UI_DEPS_FLAG_FILE)
ui-test: $(UI_DEPS_FLAG_FILE)
$(NPM_BIN) --prefix awx/ui run test
ui-lint: $(UI_DEPS_FLAG_FILE)
$(NPM_BIN) run --prefix awx/ui jshint
$(NPM_BIN) run --prefix awx/ui lint
# A standard go-to target for API developers to use building the frontend
ui: clean-ui ui-devel
@@ -565,22 +551,22 @@ docker-auth:
fi;
# Docker isolated rampart
docker-isolated:
TAG=$(COMPOSE_TAG) DEV_DOCKER_TAG_BASE=$(DEV_DOCKER_TAG_BASE) docker-compose -f tools/docker-compose.yml -f tools/docker-isolated-override.yml create
docker start tools_awx_1
docker start tools_isolated_1
echo "__version__ = '`git describe --long | cut -d - -f 1-1`'" | docker exec -i tools_isolated_1 /bin/bash -c "cat > /venv/awx/lib/python$(PYTHON_VERSION)/site-packages/awx.py"
docker-compose-isolated:
CURRENT_UID=$(shell id -u) TAG=$(COMPOSE_TAG) DEV_DOCKER_TAG_BASE=$(DEV_DOCKER_TAG_BASE) docker-compose -f tools/docker-compose.yml -f tools/docker-isolated-override.yml up
# Docker Compose Development environment
docker-compose: docker-auth
CURRENT_UID=$(shell id -u) TAG=$(COMPOSE_TAG) DEV_DOCKER_TAG_BASE=$(DEV_DOCKER_TAG_BASE) docker-compose -f tools/docker-compose.yml up --no-recreate awx
CURRENT_UID=$(shell id -u) OS="$(shell docker info | grep 'Operating System')" TAG=$(COMPOSE_TAG) DEV_DOCKER_TAG_BASE=$(DEV_DOCKER_TAG_BASE) docker-compose -f tools/docker-compose.yml up --no-recreate awx
docker-compose-cluster: docker-auth
CURRENT_UID=$(shell id -u) TAG=$(COMPOSE_TAG) DEV_DOCKER_TAG_BASE=$(DEV_DOCKER_TAG_BASE) docker-compose -f tools/docker-compose-cluster.yml up
docker-compose-credential-plugins: docker-auth
echo -e "\033[0;31mTo generate a CyberArk Conjur API key: docker exec -it tools_conjur_1 conjurctl account create quick-start\033[0m"
CURRENT_UID=$(shell id -u) TAG=$(COMPOSE_TAG) DEV_DOCKER_TAG_BASE=$(DEV_DOCKER_TAG_BASE) docker-compose -f tools/docker-compose.yml -f tools/docker-credential-plugins-override.yml up --no-recreate awx
docker-compose-test: docker-auth
cd tools && CURRENT_UID=$(shell id -u) TAG=$(COMPOSE_TAG) DEV_DOCKER_TAG_BASE=$(DEV_DOCKER_TAG_BASE) docker-compose run --rm --service-ports awx /bin/bash
cd tools && CURRENT_UID=$(shell id -u) OS="$(shell docker info | grep 'Operating System')" TAG=$(COMPOSE_TAG) DEV_DOCKER_TAG_BASE=$(DEV_DOCKER_TAG_BASE) docker-compose run --rm --service-ports awx /bin/bash
docker-compose-runtest:
cd tools && CURRENT_UID=$(shell id -u) TAG=$(COMPOSE_TAG) DEV_DOCKER_TAG_BASE=$(DEV_DOCKER_TAG_BASE) docker-compose run --rm --service-ports awx /start_tests.sh
@@ -588,12 +574,7 @@ docker-compose-runtest:
docker-compose-build-swagger:
cd tools && CURRENT_UID=$(shell id -u) TAG=$(COMPOSE_TAG) DEV_DOCKER_TAG_BASE=$(DEV_DOCKER_TAG_BASE) docker-compose run --rm --service-ports awx /start_tests.sh swagger
docker-compose-genschema:
cd tools && CURRENT_UID=$(shell id -u) TAG=$(COMPOSE_TAG) DEV_DOCKER_TAG_BASE=$(DEV_DOCKER_TAG_BASE) docker-compose run --rm --service-ports awx /start_tests.sh genschema
mv swagger.json schema.json
docker-compose-detect-schema-change:
$(MAKE) docker-compose-genschema
detect-schema-change: genschema
curl https://s3.amazonaws.com/awx-public-ci-files/schema.json -o reference-schema.json
# Ignore differences in whitespace with -b
diff -u -b reference-schema.json schema.json
@@ -606,12 +587,14 @@ docker-compose-build: awx-devel-build
# Base development image build
awx-devel-build:
docker build -t ansible/awx_devel -f tools/docker-compose/Dockerfile .
docker build -t ansible/awx_devel -f tools/docker-compose/Dockerfile \
--cache-from=$(DEV_DOCKER_TAG_BASE)/awx_devel:devel \
--cache-from=$(DEV_DOCKER_TAG_BASE)/awx_devel:$(COMPOSE_TAG) .
docker tag ansible/awx_devel $(DEV_DOCKER_TAG_BASE)/awx_devel:$(COMPOSE_TAG)
#docker push $(DEV_DOCKER_TAG_BASE)/awx_devel:$(COMPOSE_TAG)
# For use when developing on "isolated" AWX deployments
awx-isolated-build:
docker-compose-isolated-build: awx-devel-build
docker build -t ansible/awx_isolated -f tools/docker-isolated/Dockerfile .
docker tag ansible/awx_isolated $(DEV_DOCKER_TAG_BASE)/awx_isolated:$(COMPOSE_TAG)
#docker push $(DEV_DOCKER_TAG_BASE)/awx_isolated:$(COMPOSE_TAG)
@@ -631,6 +614,9 @@ docker-compose-elk: docker-auth
docker-compose-cluster-elk: docker-auth
TAG=$(COMPOSE_TAG) DEV_DOCKER_TAG_BASE=$(DEV_DOCKER_TAG_BASE) docker-compose -f tools/docker-compose-cluster.yml -f tools/elastic/docker-compose.logstash-link-cluster.yml -f tools/elastic/docker-compose.elastic-override.yml up --no-recreate
prometheus:
docker run -u0 --net=tools_default --link=`docker ps | egrep -o "tools_awx(_run)?_([^ ]+)?"`:awxweb --volume `pwd`/tools/prometheus:/prometheus --name prometheus -d -p 0.0.0.0:9090:9090 prom/prometheus --web.enable-lifecycle --config.file=/prometheus/prometheus.yml
minishift-dev:
ansible-playbook -i localhost, -e devtree_directory=$(CURDIR) tools/clusterdevel/start_minishift_dev.yml

View File

@@ -1 +1 @@
3.0.1
5.0.0

View File

@@ -37,9 +37,9 @@ if HAS_DJANGO is True:
# This line exists to make sure we don't regress on FIPS support if we
# upgrade Django; if you're upgrading Django and see this error,
# update the version check below, and confirm that FIPS still works.
if django.__version__ != '1.11.16':
raise RuntimeError("Django version other than 1.11.16 detected {}. \
Subclassing BaseDatabaseSchemaEditor is known to work for Django 1.11.16 \
if django.__version__ != '1.11.20':
raise RuntimeError("Django version other than 1.11.20 detected {}. \
Subclassing BaseDatabaseSchemaEditor is known to work for Django 1.11.20 \
and may not work in newer Django versions.".format(django.__version__))

View File

@@ -101,6 +101,10 @@ class DeprecatedCredentialField(serializers.IntegerField):
super(DeprecatedCredentialField, self).__init__(**kwargs)
def to_internal_value(self, pk):
try:
pk = int(pk)
except ValueError:
self.fail('invalid')
try:
Credential.objects.get(pk=pk)
except ObjectDoesNotExist:

View File

@@ -24,20 +24,6 @@ from rest_framework.filters import BaseFilterBackend
# AWX
from awx.main.utils import get_type_for_model, to_python_boolean
from awx.main.utils.db import get_all_field_names
from awx.main.models.credential import CredentialType
class V1CredentialFilterBackend(BaseFilterBackend):
'''
For /api/v1/ requests, filter out v2 (custom) credentials
'''
def filter_queryset(self, request, queryset, view):
# TODO: remove in 3.3
from awx.api.versioning import get_request_version
if get_request_version(request) == 1:
queryset = queryset.filter(credential_type__managed_by_tower=True)
return queryset
class TypeFilterBackend(BaseFilterBackend):
@@ -223,7 +209,7 @@ class FieldLookupBackend(BaseFilterBackend):
raise ValueError('%s is not searchable' % new_lookup[:-8])
new_lookups = []
for rm_field in related_model._meta.fields:
if rm_field.name in ('username', 'first_name', 'last_name', 'email', 'name', 'description'):
if rm_field.name in ('username', 'first_name', 'last_name', 'email', 'name', 'description', 'playbook'):
new_lookups.append('{}__{}__icontains'.format(new_lookup[:-8], rm_field.name))
return value, new_lookups
else:
@@ -292,39 +278,6 @@ class FieldLookupBackend(BaseFilterBackend):
key = key[5:]
q_not = True
# Make legacy v1 Job/Template fields work for backwards compatability
# TODO: remove after API v1 deprecation period
if queryset.model._meta.object_name in ('JobTemplate', 'Job') and key in (
'credential', 'vault_credential', 'cloud_credential', 'network_credential'
) or queryset.model._meta.object_name in ('InventorySource', 'InventoryUpdate') and key == 'credential':
key = 'credentials'
# Make legacy v1 Credential fields work for backwards compatability
# TODO: remove after API v1 deprecation period
#
# convert v1 `Credential.kind` queries to `Credential.credential_type__pk`
if queryset.model._meta.object_name == 'Credential' and key == 'kind':
key = key.replace('kind', 'credential_type')
if 'ssh' in values:
# In 3.2, SSH and Vault became separate credential types, but in the v1 API,
# they're both still "kind=ssh"
# under the hood, convert `/api/v1/credentials/?kind=ssh` to
# `/api/v1/credentials/?or__credential_type=<ssh_pk>&or__credential_type=<vault_pk>`
values = set(values)
values.add('vault')
values = list(values)
q_or = True
for i, kind in enumerate(values):
if kind == 'vault':
type_ = CredentialType.objects.get(kind=kind)
else:
type_ = CredentialType.from_v1_kind(kind)
if type_ is None:
raise ParseError(_('cannot filter on kind %s') % kind)
values[i] = type_.pk
# Convert value(s) to python and add to the appropriate list.
for value in values:
if q_int:
@@ -402,6 +355,8 @@ class OrderByBackend(BaseFilterBackend):
order_by = value.split(',')
else:
order_by = (value,)
if order_by is None:
order_by = self.get_default_ordering(view)
if order_by:
order_by = self._validate_ordering_fields(queryset.model, order_by)
@@ -428,6 +383,12 @@ class OrderByBackend(BaseFilterBackend):
# Return a 400 for invalid field names.
raise ParseError(*e.args)
def get_default_ordering(self, view):
ordering = getattr(view, 'ordering', None)
if isinstance(ordering, str):
return (ordering,)
return ordering
def _validate_ordering_fields(self, model, order_by):
for field_name in order_by:
# strip off the negation prefix `-` if it exists

View File

@@ -33,12 +33,20 @@ from rest_framework.negotiation import DefaultContentNegotiation
# AWX
from awx.api.filters import FieldLookupBackend
from awx.main.models import * # noqa
from awx.main.models import (
UnifiedJob, UnifiedJobTemplate, User, Role, Credential
)
from awx.main.access import access_registry
from awx.main.utils import * # noqa
from awx.main.utils import (
camelcase_to_underscore,
get_search_fields,
getattrd,
get_object_or_400,
decrypt_field
)
from awx.main.utils.db import get_all_field_names
from awx.api.serializers import ResourceAccessListElementSerializer, CopySerializer, UserSerializer
from awx.api.versioning import URLPathVersioning, get_request_version
from awx.api.versioning import URLPathVersioning
from awx.api.metadata import SublistAttachDetatchMetadata, Metadata
__all__ = ['APIView', 'GenericAPIView', 'ListAPIView', 'SimpleListAPIView',
@@ -92,6 +100,8 @@ class LoggedLoginView(auth_views.LoginView):
return ret
else:
if 'username' in self.request.POST:
logger.warn(smart_text(u"Login failed for user {} from {}".format(self.request.POST.get('username'),request.META.get('REMOTE_ADDR', None))))
ret.status_code = 401
return ret
@@ -109,39 +119,12 @@ class LoggedLogoutView(auth_views.LogoutView):
return ret
def get_view_name(cls, suffix=None):
'''
Wrapper around REST framework get_view_name() to support get_name() method
and view_name property on a view class.
'''
name = ''
if hasattr(cls, 'get_name') and callable(cls.get_name):
name = cls().get_name()
elif hasattr(cls, 'view_name'):
if callable(cls.view_name):
name = cls.view_name()
else:
name = cls.view_name
if name:
return ('%s %s' % (name, suffix)) if suffix else name
return views.get_view_name(cls, suffix=None)
def get_view_description(view, html=False):
'''Wrapper around REST framework get_view_description() to continue
to support our historical div.
def get_view_description(cls, request, html=False):
'''
Wrapper around REST framework get_view_description() to support
get_description() method and view_description property on a view class.
'''
if hasattr(cls, 'get_description') and callable(cls.get_description):
desc = cls().get_description(request, html=html)
cls = type(cls.__name__, (object,), {'__doc__': desc})
elif hasattr(cls, 'view_description'):
if callable(cls.view_description):
view_desc = cls.view_description()
else:
view_desc = cls.view_description
cls = type(cls.__name__, (object,), {'__doc__': view_desc})
desc = views.get_view_description(cls, html=html)
desc = views.get_view_description(view, html=html)
if html:
desc = '<div class="description">%s</div>' % desc
return mark_safe(desc)
@@ -254,14 +237,6 @@ class APIView(views.APIView):
# `curl https://user:pass@tower.example.org/api/v2/job_templates/N/launch/`
return 'Bearer realm=api authorization_url=/api/o/authorize/'
def get_view_description(self, html=False):
"""
Return some descriptive text for the view, as used in OPTIONS responses
and in the browsable API.
"""
func = self.settings.VIEW_DESCRIPTION_FUNCTION
return func(self.__class__, getattr(self, '_request', None), html)
def get_description_context(self):
return {
'view': self,
@@ -270,20 +245,14 @@ class APIView(views.APIView):
'swagger_method': getattr(self.request, 'swagger_method', None),
}
def get_description(self, request, html=False):
self.request = request
@property
def description(self):
template_list = []
for klass in inspect.getmro(type(self)):
template_basename = camelcase_to_underscore(klass.__name__)
template_list.append('api/%s.md' % template_basename)
context = self.get_description_context()
# "v2" -> 2
default_version = int(settings.REST_FRAMEWORK['DEFAULT_VERSION'].lstrip('v'))
request_version = get_request_version(self.request)
if request_version is not None and request_version < default_version:
context['deprecated'] = True
description = render_to_string(template_list, context)
if context.get('deprecated') and context.get('swagger_method') is None:
# render deprecation messages at the very top
@@ -379,12 +348,14 @@ class GenericAPIView(generics.GenericAPIView, APIView):
'model_verbose_name_plural': smart_text(self.model._meta.verbose_name_plural),
})
serializer = self.get_serializer()
metadata = self.metadata_class()
metadata.request = self.request
for method, key in [
('GET', 'serializer_fields'),
('POST', 'serializer_create_fields'),
('PUT', 'serializer_update_fields')
]:
d[key] = self.metadata_class().get_serializer_info(serializer, method=method)
d[key] = metadata.get_serializer_info(serializer, method=method)
d['settings'] = settings
return d
@@ -805,6 +776,7 @@ class RetrieveUpdateDestroyAPIView(RetrieveUpdateAPIView, DestroyAPIView):
class ResourceAccessList(ParentMixin, ListAPIView):
serializer_class = ResourceAccessListElementSerializer
ordering = ('username',)
def get_queryset(self):
obj = self.get_parent_object()
@@ -831,10 +803,6 @@ class CopyAPIView(GenericAPIView):
new_in_330 = True
new_in_api_v2 = True
def v1_not_allowed(self):
return Response({'detail': 'Action only possible starting with v2 API.'},
status=status.HTTP_404_NOT_FOUND)
def _get_copy_return_serializer(self, *args, **kwargs):
if not self.copy_return_serializer_class:
return self.get_serializer(*args, **kwargs)
@@ -848,15 +816,15 @@ class CopyAPIView(GenericAPIView):
def _decrypt_model_field_if_needed(obj, field_name, field_val):
if field_name in getattr(type(obj), 'REENCRYPTION_BLACKLIST_AT_COPY', []):
return field_val
if isinstance(field_val, dict):
if isinstance(obj, Credential) and field_name == 'inputs':
for secret in obj.credential_type.secret_fields:
if secret in field_val:
field_val[secret] = decrypt_field(obj, secret)
elif isinstance(field_val, dict):
for sub_field in field_val:
if isinstance(sub_field, str) \
and isinstance(field_val[sub_field], str):
try:
field_val[sub_field] = decrypt_field(obj, field_name, sub_field)
except AttributeError:
# Catching the corner case with v1 credential fields
field_val[sub_field] = decrypt_field(obj, sub_field)
field_val[sub_field] = decrypt_field(obj, field_name, sub_field)
elif isinstance(field_val, str):
try:
field_val = decrypt_field(obj, field_name)
@@ -941,21 +909,20 @@ class CopyAPIView(GenericAPIView):
return ret
def get(self, request, *args, **kwargs):
if get_request_version(request) < 2:
return self.v1_not_allowed()
obj = self.get_object()
if not request.user.can_access(obj.__class__, 'read', obj):
raise PermissionDenied()
create_kwargs = self._build_create_dict(obj)
for key in create_kwargs:
create_kwargs[key] = getattr(create_kwargs[key], 'pk', None) or create_kwargs[key]
can_copy = request.user.can_access(self.model, 'add', create_kwargs) and \
request.user.can_access(self.model, 'copy_related', obj)
try:
can_copy = request.user.can_access(self.model, 'add', create_kwargs) and \
request.user.can_access(self.model, 'copy_related', obj)
except PermissionDenied:
return Response({'can_copy': False})
return Response({'can_copy': can_copy})
def post(self, request, *args, **kwargs):
if get_request_version(request) < 2:
return self.v1_not_allowed()
obj = self.get_object()
create_kwargs = self._build_create_dict(obj)
create_kwargs_check = {}
@@ -972,7 +939,7 @@ class CopyAPIView(GenericAPIView):
None, None, self.model, obj, request.user, create_kwargs=create_kwargs,
copy_name=serializer.validated_data.get('name', '')
)
if hasattr(new_obj, 'admin_role') and request.user not in new_obj.admin_role:
if hasattr(new_obj, 'admin_role') and request.user not in new_obj.admin_role.members.all():
new_obj.admin_role.members.add(request.user)
if sub_objs:
permission_check_func = None

View File

@@ -232,28 +232,13 @@ class RoleMetadata(Metadata):
return metadata
# TODO: Tower 3.3 remove class and all uses in views.py when API v1 is removed
class JobTypeMetadata(Metadata):
def get_field_info(self, field):
res = super(JobTypeMetadata, self).get_field_info(field)
if field.field_name == 'job_type':
index = 0
for choice in res['choices']:
if choice[0] == 'scan':
res['choices'].pop(index)
break
index += 1
return res
class SublistAttachDetatchMetadata(Metadata):
def determine_actions(self, request, view):
actions = super(SublistAttachDetatchMetadata, self).determine_actions(request, view)
method = 'POST'
if method in actions:
for field in actions[method]:
for field in list(actions[method].keys()):
if field == 'id':
continue
actions[method].pop(field)

15
awx/api/metrics.py Normal file
View File

@@ -0,0 +1,15 @@
# Copyright (c) 2017 Ansible, Inc.
# All Rights Reserved.
from django.conf.urls import url
from awx.api.views import (
MetricsView
)
urls = [
url(r'^$', MetricsView.as_view(), name='metrics_view'),
]
__all__ = ['urls']

View File

@@ -18,7 +18,7 @@ class Pagination(pagination.PageNumberPagination):
url = self.request and self.request.get_full_path() or ''
url = url.encode('utf-8')
page_number = self.page.next_page_number()
return replace_query_param(url, self.page_query_param, page_number)
return replace_query_param(self.cap_page_size(url), self.page_query_param, page_number)
def get_previous_link(self):
if not self.page.has_previous():
@@ -26,4 +26,16 @@ class Pagination(pagination.PageNumberPagination):
url = self.request and self.request.get_full_path() or ''
url = url.encode('utf-8')
page_number = self.page.previous_page_number()
return replace_query_param(url, self.page_query_param, page_number)
return replace_query_param(self.cap_page_size(url), self.page_query_param, page_number)
def cap_page_size(self, url):
if int(self.request.query_params.get(self.page_size_query_param, 0)) > self.max_page_size:
url = replace_query_param(url, self.page_size_query_param, self.max_page_size)
return url
def get_html_context(self):
context = super().get_html_context()
context['page_links'] = [pl._replace(url=self.cap_page_size(pl.url))
for pl in context['page_links']]
return context

View File

@@ -9,13 +9,13 @@ from rest_framework.exceptions import MethodNotAllowed, PermissionDenied
from rest_framework import permissions
# AWX
from awx.main.access import * # noqa
from awx.main.models import * # noqa
from awx.main.access import check_user_access
from awx.main.models import Inventory, UnifiedJob
from awx.main.utils import get_object_or_400
logger = logging.getLogger('awx.api.permissions')
__all__ = ['ModelAccessPermission', 'JobTemplateCallbackPermission',
__all__ = ['ModelAccessPermission', 'JobTemplateCallbackPermission', 'VariableDataPermission',
'TaskPermission', 'ProjectUpdatePermission', 'InventoryInventorySourcesUpdatePermission',
'UserPermission', 'IsSuperUser', 'InstanceGroupTowerPermission',]
@@ -74,12 +74,8 @@ class ModelAccessPermission(permissions.BasePermission):
# FIXME: For some reason this needs to return True
# because it is first called with obj=None?
return True
if getattr(view, 'is_variable_data', False):
return check_user_access(request.user, view.model, 'change', obj,
dict(variables=request.data))
else:
return check_user_access(request.user, view.model, 'change', obj,
request.data)
return check_user_access(request.user, view.model, 'change', obj,
request.data)
def check_patch_permissions(self, request, view, obj=None):
return self.check_put_permissions(request, view, obj)
@@ -103,8 +99,7 @@ class ModelAccessPermission(permissions.BasePermission):
return False
# Always allow superusers
if getattr(view, 'always_allow_superuser', True) and request.user.is_superuser \
and not hasattr(request.user, 'oauth_scopes'):
if getattr(view, 'always_allow_superuser', True) and request.user.is_superuser:
return True
# Check if view supports the request method before checking permission
@@ -164,6 +159,15 @@ class JobTemplateCallbackPermission(ModelAccessPermission):
return True
class VariableDataPermission(ModelAccessPermission):
def check_put_permissions(self, request, view, obj=None):
if not obj:
return True
return check_user_access(request.user, view.model, 'change', obj,
dict(variables=request.data))
class TaskPermission(ModelAccessPermission):
'''
Permission checks used for API callbacks from running a task.

File diff suppressed because it is too large Load Diff

View File

@@ -13,6 +13,17 @@ from rest_framework.views import APIView
from rest_framework_swagger import renderers
class SuperUserSchemaGenerator(SchemaGenerator):
def has_view_permissions(self, path, method, view):
#
# Generate the Swagger schema as if you were a superuser and
# permissions didn't matter; this short-circuits the schema path
# discovery to include _all_ potential paths in the API.
#
return True
class AutoSchema(DRFAuthSchema):
def get_link(self, path, method, base_url):
@@ -42,7 +53,6 @@ class AutoSchema(DRFAuthSchema):
return link
def get_description(self, path, method):
self.view._request = self.view.request
setattr(self.view.request, 'swagger_method', method)
description = super(AutoSchema, self).get_description(path, method)
return description
@@ -59,7 +69,7 @@ class SwaggerSchemaView(APIView):
]
def get(self, request):
generator = SchemaGenerator(
generator = SuperUserSchemaGenerator(
title='Ansible Tower API',
patterns=None,
urlconf=None

View File

@@ -5,7 +5,7 @@ The following lists the expected format and details of our rrules:
* INTERVAL is required
* SECONDLY is not supported
* TZID is not supported
* RRULE must preceed the rule statements
* RRULE must precede the rule statements
* BYDAY is supported but not BYDAY with a numerical prefix
* BYYEARDAY and BYWEEKNO are not supported
* Only one rrule statement per schedule is supported

View File

@@ -29,17 +29,6 @@ to the redirect_uri specified in the application. The client application will th
AWX will respond with the `access_token`, `token_type`, `refresh_token`, and `expires_in`. For more
information on testing this flow, refer to [django-oauth-toolkit](http://django-oauth-toolkit.readthedocs.io/en/latest/tutorial/tutorial_01.html#test-your-authorization-server).
## Create Token for an Application using Implicit grant type
Suppose we have an application "admin's app" of grant type `implicit`.
In API browser, first make sure the user is logged in via session auth, then visit authorization
endpoint with given parameters:
```text
http://localhost:8013/api/o/authorize/?response_type=token&client_id=L0uQQWW8pKX51hoqIRQGsuqmIdPi2AcXZ9EJRGmj&scope=read
```
Here the value of `client_id` should be the same as that of `client_id` field of underlying application.
On success, an authorization page should be displayed asking the logged in user to grant/deny the access token.
Once the user clicks on 'grant', the API browser will try POSTing to the same endpoint with the same parameters
in POST body, on success a 302 redirect will be returned.
## Create Token for an Application using Password grant type
@@ -56,6 +45,7 @@ For example:
```bash
curl -X POST \
-H "Content-Type: application/x-www-form-urlencoded" \
-d "grant_type=password&username=<username>&password=<password>&scope=read" \
-u "gwSPoasWSdNkMDtBN3Hu2WYQpPWCO9SwUEsKK22l:fI6ZpfocHYBGfm1tP92r0yIgCyfRdDQt0Tos9L8a4fNsJjQQMwp9569e
IaUBsaVDgt2eiwOGe0bg5m5vCSstClZmtdy359RVx2rQK5YlIWyPlrolpt2LEpVeKXWaiybo" \
@@ -85,6 +75,7 @@ format:
The `/api/o/token/` endpoint is used for refreshing access token:
```bash
curl -X POST \
-H "Content-Type: application/x-www-form-urlencoded" \
-d "grant_type=refresh_token&refresh_token=AL0NK9TTpv0qp54dGbC4VUZtsZ9r8z" \
-u "gwSPoasWSdNkMDtBN3Hu2WYQpPWCO9SwUEsKK22l:fI6ZpfocHYBGfm1tP92r0yIgCyfRdDQt0Tos9L8a4fNsJjQQMwp9569eIaUBsaVDgt2eiwOGe0bg5m5vCSstClZmtdy359RVx2rQK5YlIWyPlrolpt2LEpVeKXWaiybo" \
http://localhost:8013/api/o/token/ -i
@@ -114,6 +105,7 @@ Revoking is done by POSTing to `/api/o/revoke_token/` with the token to revoke a
```bash
curl -X POST -d "token=rQONsve372fQwuc2pn76k3IHDCYpi7" \
-H "Content-Type: application/x-www-form-urlencoded" \
-u "gwSPoasWSdNkMDtBN3Hu2WYQpPWCO9SwUEsKK22l:fI6ZpfocHYBGfm1tP92r0yIgCyfRdDQt0Tos9L8a4fNsJjQQMwp9569eIaUBsaVDgt2eiwOGe0bg5m5vCSstClZmtdy359RVx2rQK5YlIWyPlrolpt2LEpVeKXWaiybo" \
http://localhost:8013/api/o/revoke_token/ -i
```

View File

@@ -8,15 +8,15 @@ job template.
For example, using curl:
curl -H "Content-Type: application/json" -d '{"host_config_key": "HOST_CONFIG_KEY"}' http://server/api/v1/job_templates/N/callback/
curl -H "Content-Type: application/json" -d '{"host_config_key": "HOST_CONFIG_KEY"}' http://server/api/v2/job_templates/N/callback/
Or using wget:
wget -O /dev/null --post-data='{"host_config_key": "HOST_CONFIG_KEY"}' --header=Content-Type:application/json http://server/api/v1/job_templates/N/callback/
wget -O /dev/null --post-data='{"host_config_key": "HOST_CONFIG_KEY"}' --header=Content-Type:application/json http://server/api/v2/job_templates/N/callback/
You may also pass `extra_vars` to the callback:
curl -H "Content-Type: application/json" -d '{"host_config_key": "HOST_CONFIG_KEY", "extra_vars": {"key": "value"}}' http://server/api/v1/job_templates/N/callback/
curl -H "Content-Type: application/json" -d '{"host_config_key": "HOST_CONFIG_KEY", "extra_vars": {"key": "value"}}' http://server/api/v2/job_templates/N/callback/
The response will return status 202 if the request is valid, 403 for an
invalid host config key, or 400 if the host cannot be determined from the
@@ -30,7 +30,7 @@ A GET request may be used to verify that the correct host will be selected.
This request must authenticate as a valid user with permission to edit the
job template. For example:
curl http://user:password@server/api/v1/job_templates/N/callback/
curl http://user:password@server/api/v2/job_templates/N/callback/
The response will include the host config key as well as the host name(s)
that would match the request:

View File

@@ -3,7 +3,7 @@ Launch a Job Template:
Make a POST request to this resource to launch the system job template.
Variables specified inside of the parameter `extra_vars` are passed to the
system job task as command line parameters. These tasks can be ran manually
system job task as command line parameters. These tasks can be run manually
on the host system via the `awx-manage` command.
For example on `cleanup_jobs` and `cleanup_activitystream`:

View File

@@ -6,4 +6,4 @@ One result should be returned containing the following fields:
{% include "api/_result_fields_common.md" %}
Use the primary URL for the user (/api/v1/users/N/) to modify the user.
Use the primary URL for the user (/api/v2/users/N/) to modify the user.

View File

@@ -4,4 +4,7 @@
from __future__ import absolute_import, unicode_literals
from .urls import urlpatterns
__all__ = ['urlpatterns']
__all__ = ['urlpatterns', 'app_name']
app_name = 'api'

View File

@@ -12,6 +12,8 @@ from awx.api.views import (
CredentialOwnerUsersList,
CredentialOwnerTeamsList,
CredentialCopy,
CredentialInputSourceSubList,
CredentialExternalTest,
)
@@ -24,6 +26,8 @@ urls = [
url(r'^(?P<pk>[0-9]+)/owner_users/$', CredentialOwnerUsersList.as_view(), name='credential_owner_users_list'),
url(r'^(?P<pk>[0-9]+)/owner_teams/$', CredentialOwnerTeamsList.as_view(), name='credential_owner_teams_list'),
url(r'^(?P<pk>[0-9]+)/copy/$', CredentialCopy.as_view(), name='credential_copy'),
url(r'^(?P<pk>[0-9]+)/input_sources/$', CredentialInputSourceSubList.as_view(), name='credential_input_source_sublist'),
url(r'^(?P<pk>[0-9]+)/test/$', CredentialExternalTest.as_view(), name='credential_external_test'),
]
__all__ = ['urls']

View File

@@ -0,0 +1,17 @@
# Copyright (c) 2019 Ansible, Inc.
# All Rights Reserved.
from django.conf.urls import url
from awx.api.views import (
CredentialInputSourceDetail,
CredentialInputSourceList,
)
urls = [
url(r'^$', CredentialInputSourceList.as_view(), name='credential_input_source_list'),
url(r'^(?P<pk>[0-9]+)/$', CredentialInputSourceDetail.as_view(), name='credential_input_source_detail'),
]
__all__ = ['urls']

View File

@@ -8,6 +8,7 @@ from awx.api.views import (
CredentialTypeDetail,
CredentialTypeCredentialList,
CredentialTypeActivityStreamList,
CredentialTypeExternalTest,
)
@@ -16,6 +17,7 @@ urls = [
url(r'^(?P<pk>[0-9]+)/$', CredentialTypeDetail.as_view(), name='credential_type_detail'),
url(r'^(?P<pk>[0-9]+)/credentials/$', CredentialTypeCredentialList.as_view(), name='credential_type_credential_list'),
url(r'^(?P<pk>[0-9]+)/activity_stream/$', CredentialTypeActivityStreamList.as_view(), name='credential_type_activity_stream_list'),
url(r'^(?P<pk>[0-9]+)/test/$', CredentialTypeExternalTest.as_view(), name='credential_type_external_test'),
]
__all__ = ['urls']

View File

@@ -16,8 +16,6 @@ from awx.api.views import (
HostSmartInventoriesList,
HostAdHocCommandsList,
HostAdHocCommandEventsList,
HostFactVersionsList,
HostFactCompareView,
HostInsights,
)
@@ -35,8 +33,6 @@ urls = [
url(r'^(?P<pk>[0-9]+)/smart_inventories/$', HostSmartInventoriesList.as_view(), name='host_smart_inventories_list'),
url(r'^(?P<pk>[0-9]+)/ad_hoc_commands/$', HostAdHocCommandsList.as_view(), name='host_ad_hoc_commands_list'),
url(r'^(?P<pk>[0-9]+)/ad_hoc_command_events/$', HostAdHocCommandEventsList.as_view(), name='host_ad_hoc_command_events_list'),
url(r'^(?P<pk>[0-9]+)/fact_versions/$', HostFactVersionsList.as_view(), name='host_fact_versions_list'),
url(r'^(?P<pk>[0-9]+)/fact_view/$', HostFactCompareView.as_view(), name='host_fact_compare_view'),
url(r'^(?P<pk>[0-9]+)/insights/$', HostInsights.as_view(), name='host_insights'),
]

View File

@@ -6,7 +6,6 @@ from django.conf.urls import url
from awx.api.views import (
JobList,
JobDetail,
JobStart,
JobCancel,
JobRelaunch,
JobCreateSchedule,
@@ -23,7 +22,6 @@ from awx.api.views import (
urls = [
url(r'^$', JobList.as_view(), name='job_list'),
url(r'^(?P<pk>[0-9]+)/$', JobDetail.as_view(), name='job_detail'),
url(r'^(?P<pk>[0-9]+)/start/$', JobStart.as_view(), name='job_start'), # Todo: Remove In 3.3
url(r'^(?P<pk>[0-9]+)/cancel/$', JobCancel.as_view(), name='job_cancel'),
url(r'^(?P<pk>[0-9]+)/relaunch/$', JobRelaunch.as_view(), name='job_relaunch'),
url(r'^(?P<pk>[0-9]+)/create_schedule/$', JobCreateSchedule.as_view(), name='job_create_schedule'),

View File

@@ -11,10 +11,9 @@ from awx.api.generics import (
)
from awx.api.views import (
ApiRootView,
ApiV1RootView,
ApiV2RootView,
ApiV1PingView,
ApiV1ConfigView,
ApiV2PingView,
ApiV2ConfigView,
AuthView,
UserMeList,
DashboardView,
@@ -34,6 +33,8 @@ from awx.api.views import (
OAuth2ApplicationDetail,
)
from awx.api.views.metrics import MetricsView
from .organization import urls as organization_urls
from .user import urls as user_urls
from .project import urls as project_urls
@@ -47,6 +48,7 @@ from .inventory_update import urls as inventory_update_urls
from .inventory_script import urls as inventory_script_urls
from .credential_type import urls as credential_type_urls
from .credential import urls as credential_urls
from .credential_input_source import urls as credential_input_source_urls
from .role import urls as role_urls
from .job_template import urls as job_template_urls
from .job import urls as job_urls
@@ -71,10 +73,25 @@ from .oauth2 import urls as oauth2_urls
from .oauth2_root import urls as oauth2_root_urls
v1_urls = [
url(r'^$', ApiV1RootView.as_view(), name='api_v1_root_view'),
url(r'^ping/$', ApiV1PingView.as_view(), name='api_v1_ping_view'),
url(r'^config/$', ApiV1ConfigView.as_view(), name='api_v1_config_view'),
v2_urls = [
url(r'^$', ApiV2RootView.as_view(), name='api_v2_root_view'),
url(r'^credential_types/', include(credential_type_urls)),
url(r'^credential_input_sources/', include(credential_input_source_urls)),
url(r'^hosts/(?P<pk>[0-9]+)/ansible_facts/$', HostAnsibleFactsDetail.as_view(), name='host_ansible_facts_detail'),
url(r'^jobs/(?P<pk>[0-9]+)/extra_credentials/$', JobExtraCredentialsList.as_view(), name='job_extra_credentials_list'),
url(r'^jobs/(?P<pk>[0-9]+)/credentials/$', JobCredentialsList.as_view(), name='job_credentials_list'),
url(r'^job_templates/(?P<pk>[0-9]+)/extra_credentials/$', JobTemplateExtraCredentialsList.as_view(), name='job_template_extra_credentials_list'),
url(r'^job_templates/(?P<pk>[0-9]+)/credentials/$', JobTemplateCredentialsList.as_view(), name='job_template_credentials_list'),
url(r'^schedules/preview/$', SchedulePreview.as_view(), name='schedule_rrule'),
url(r'^schedules/zoneinfo/$', ScheduleZoneInfo.as_view(), name='schedule_zoneinfo'),
url(r'^applications/$', OAuth2ApplicationList.as_view(), name='o_auth2_application_list'),
url(r'^applications/(?P<pk>[0-9]+)/$', OAuth2ApplicationDetail.as_view(), name='o_auth2_application_detail'),
url(r'^applications/(?P<pk>[0-9]+)/tokens/$', ApplicationOAuth2TokenList.as_view(), name='application_o_auth2_token_list'),
url(r'^tokens/$', OAuth2TokenList.as_view(), name='o_auth2_token_list'),
url(r'^', include(oauth2_urls)),
url(r'^metrics/$', MetricsView.as_view(), name='metrics_view'),
url(r'^ping/$', ApiV2PingView.as_view(), name='api_v2_ping_view'),
url(r'^config/$', ApiV2ConfigView.as_view(), name='api_v2_config_view'),
url(r'^auth/$', AuthView.as_view()),
url(r'^me/$', UserMeList.as_view(), name='user_me_list'),
url(r'^dashboard/$', DashboardView.as_view(), name='dashboard_view'),
@@ -116,28 +133,10 @@ v1_urls = [
url(r'^activity_stream/', include(activity_stream_urls)),
]
v2_urls = [
url(r'^$', ApiV2RootView.as_view(), name='api_v2_root_view'),
url(r'^credential_types/', include(credential_type_urls)),
url(r'^hosts/(?P<pk>[0-9]+)/ansible_facts/$', HostAnsibleFactsDetail.as_view(), name='host_ansible_facts_detail'),
url(r'^jobs/(?P<pk>[0-9]+)/extra_credentials/$', JobExtraCredentialsList.as_view(), name='job_extra_credentials_list'),
url(r'^jobs/(?P<pk>[0-9]+)/credentials/$', JobCredentialsList.as_view(), name='job_credentials_list'),
url(r'^job_templates/(?P<pk>[0-9]+)/extra_credentials/$', JobTemplateExtraCredentialsList.as_view(), name='job_template_extra_credentials_list'),
url(r'^job_templates/(?P<pk>[0-9]+)/credentials/$', JobTemplateCredentialsList.as_view(), name='job_template_credentials_list'),
url(r'^schedules/preview/$', SchedulePreview.as_view(), name='schedule_rrule'),
url(r'^schedules/zoneinfo/$', ScheduleZoneInfo.as_view(), name='schedule_zoneinfo'),
url(r'^applications/$', OAuth2ApplicationList.as_view(), name='o_auth2_application_list'),
url(r'^applications/(?P<pk>[0-9]+)/$', OAuth2ApplicationDetail.as_view(), name='o_auth2_application_detail'),
url(r'^applications/(?P<pk>[0-9]+)/tokens/$', ApplicationOAuth2TokenList.as_view(), name='application_o_auth2_token_list'),
url(r'^tokens/$', OAuth2TokenList.as_view(), name='o_auth2_token_list'),
url(r'^', include(oauth2_urls)),
]
app_name = 'api'
urlpatterns = [
url(r'^$', ApiRootView.as_view(), name='api_root_view'),
url(r'^(?P<version>(v2))/', include(v2_urls)),
url(r'^(?P<version>(v1|v2))/', include(v1_urls)),
url(r'^login/$', LoggedLoginView.as_view(
template_name='rest_framework/login.html',
extra_context={'inside_login_context': True}

View File

@@ -2,7 +2,7 @@
# All Rights Reserved.
from django.conf import settings
from django.core.urlresolvers import NoReverseMatch
from django.urls import NoReverseMatch
from rest_framework.reverse import _reverse
from rest_framework.versioning import URLPathVersioning as BaseVersioning
@@ -27,19 +27,6 @@ def drf_reverse(viewname, args=None, kwargs=None, request=None, format=None, **e
return url
def get_request_version(request):
"""
The API version of a request as an integer i.e., 1 or 2
"""
version = settings.REST_FRAMEWORK['DEFAULT_VERSION']
if request and hasattr(request, 'version'):
version = request.version
if version is None:
# For requests to /api/
return None
return int(version.lstrip('v'))
def reverse(viewname, args=None, kwargs=None, request=None, format=None, **extra):
if request is None or getattr(request, 'version', None) is None:
# We need the "current request" to determine the correct version to

File diff suppressed because it is too large Load Diff

View File

@@ -44,11 +44,9 @@ from awx.api.serializers import (
InstanceGroupSerializer,
InventoryUpdateEventSerializer,
CustomInventoryScriptSerializer,
InventoryDetailSerializer,
JobTemplateSerializer,
)
from awx.api.views.mixin import (
ActivityStreamEnforcementMixin,
RelatedJobsPreventDeleteMixin,
ControlledByScmMixin,
)
@@ -62,7 +60,7 @@ class InventoryUpdateEventsList(SubListAPIView):
serializer_class = InventoryUpdateEventSerializer
parent_model = InventoryUpdate
relationship = 'inventory_update_events'
view_name = _('Inventory Update Events List')
name = _('Inventory Update Events List')
search_fields = ('stdout',)
def finalize_response(self, request, response, *args, **kwargs):
@@ -116,17 +114,11 @@ class InventoryList(ListCreateAPIView):
model = Inventory
serializer_class = InventorySerializer
def get_queryset(self):
qs = Inventory.accessible_objects(self.request.user, 'read_role')
qs = qs.select_related('admin_role', 'read_role', 'update_role', 'use_role', 'adhoc_role')
qs = qs.prefetch_related('created_by', 'modified_by', 'organization')
return qs
class InventoryDetail(RelatedJobsPreventDeleteMixin, ControlledByScmMixin, RetrieveUpdateDestroyAPIView):
model = Inventory
serializer_class = InventoryDetailSerializer
serializer_class = InventorySerializer
def update(self, request, *args, **kwargs):
obj = self.get_object()
@@ -149,7 +141,7 @@ class InventoryDetail(RelatedJobsPreventDeleteMixin, ControlledByScmMixin, Retri
return Response(dict(error=_("{0}".format(e))), status=status.HTTP_400_BAD_REQUEST)
class InventoryActivityStreamList(ActivityStreamEnforcementMixin, SubListAPIView):
class InventoryActivityStreamList(SubListAPIView):
model = ActivityStream
serializer_class = ActivityStreamSerializer

40
awx/api/views/metrics.py Normal file
View File

@@ -0,0 +1,40 @@
# Copyright (c) 2018 Red Hat, Inc.
# All Rights Reserved.
# Python
import logging
# Django
from django.utils.translation import ugettext_lazy as _
# Django REST Framework
from rest_framework.response import Response
from rest_framework.exceptions import PermissionDenied
# AWX
# from awx.main.analytics import collectors
from awx.main.analytics.metrics import metrics
from awx.api import renderers
from awx.api.generics import (
APIView,
)
logger = logging.getLogger('awx.main.analytics')
class MetricsView(APIView):
name = _('Metrics')
swagger_topic = 'Metrics'
renderer_classes = [renderers.PlainTextRenderer,
renderers.BrowsableAPIRenderer,]
def get(self, request, format='txt'):
''' Show Metrics Details '''
if (request.user.is_superuser or request.user.is_system_auditor):
return Response(metrics().decode('UTF-8'))
raise PermissionDenied()

View File

@@ -31,48 +31,11 @@ from awx.main.models.organization import Team
from awx.main.models.projects import Project
from awx.main.models.inventory import Inventory
from awx.main.models.jobs import JobTemplate
from awx.conf.license import (
feature_enabled,
LicenseForbids,
)
from awx.api.exceptions import ActiveJobConflict
logger = logging.getLogger('awx.api.views.mixin')
class ActivityStreamEnforcementMixin(object):
'''
Mixin to check that license supports activity streams.
'''
def check_permissions(self, request):
ret = super(ActivityStreamEnforcementMixin, self).check_permissions(request)
if not feature_enabled('activity_streams'):
raise LicenseForbids(_('Your license does not allow use of the activity stream.'))
return ret
class SystemTrackingEnforcementMixin(object):
'''
Mixin to check that license supports system tracking.
'''
def check_permissions(self, request):
ret = super(SystemTrackingEnforcementMixin, self).check_permissions(request)
if not feature_enabled('system_tracking'):
raise LicenseForbids(_('Your license does not permit use of system tracking.'))
return ret
class WorkflowsEnforcementMixin(object):
'''
Mixin to check that license supports workflows.
'''
def check_permissions(self, request):
ret = super(WorkflowsEnforcementMixin, self).check_permissions(request)
if not feature_enabled('workflows') and request.method not in ('GET', 'OPTIONS', 'DELETE'):
raise LicenseForbids(_('Your license does not allow use of workflows.'))
return ret
class UnifiedJobDeletionMixin(object):
'''
Special handling when deleting a running unified job object.

View File

@@ -7,13 +7,8 @@ import logging
# Django
from django.db.models import Count
from django.contrib.contenttypes.models import ContentType
from django.utils.translation import ugettext_lazy as _
# AWX
from awx.conf.license import (
feature_enabled,
LicenseForbids,
)
from awx.main.models import (
ActivityStream,
Inventory,
@@ -50,7 +45,6 @@ from awx.api.serializers import (
InstanceGroupSerializer,
)
from awx.api.views.mixin import (
ActivityStreamEnforcementMixin,
RelatedJobsPreventDeleteMixin,
OrganizationCountsMixin,
)
@@ -69,24 +63,6 @@ class OrganizationList(OrganizationCountsMixin, ListCreateAPIView):
qs = qs.prefetch_related('created_by', 'modified_by')
return qs
def create(self, request, *args, **kwargs):
"""Create a new organzation.
If there is already an organization and the license of this
instance does not permit multiple organizations, then raise
LicenseForbids.
"""
# Sanity check: If the multiple organizations feature is disallowed
# by the license, then we are only willing to create this organization
# if no organizations exist in the system.
if (not feature_enabled('multiple_organizations') and
self.model.objects.exists()):
raise LicenseForbids(_('Your license only permits a single '
'organization to exist.'))
# Okay, create the organization as usual.
return super(OrganizationList, self).create(request, *args, **kwargs)
class OrganizationDetail(RelatedJobsPreventDeleteMixin, RetrieveUpdateDestroyAPIView):
@@ -140,6 +116,7 @@ class OrganizationUsersList(BaseUsersList):
serializer_class = UserSerializer
parent_model = Organization
relationship = 'member_role.members'
ordering = ('username',)
class OrganizationAdminsList(BaseUsersList):
@@ -148,6 +125,7 @@ class OrganizationAdminsList(BaseUsersList):
serializer_class = UserSerializer
parent_model = Organization
relationship = 'admin_role.members'
ordering = ('username',)
class OrganizationProjectsList(SubListCreateAttachDetachAPIView):
@@ -177,7 +155,7 @@ class OrganizationTeamsList(SubListCreateAttachDetachAPIView):
parent_key = 'organization'
class OrganizationActivityStreamList(ActivityStreamEnforcementMixin, SubListAPIView):
class OrganizationActivityStreamList(SubListAPIView):
model = ActivityStream
serializer_class = ActivityStreamSerializer
@@ -244,4 +222,3 @@ class OrganizationObjectRolesList(SubListAPIView):
po = self.get_parent_object()
content_type = ContentType.objects.get_for_model(self.parent_model)
return Role.objects.filter(content_type=content_type, object_id=po.pk)

View File

@@ -25,8 +25,8 @@ from awx.main.utils import (
get_custom_venv_choices,
to_python_boolean,
)
from awx.api.versioning import reverse, get_request_version, drf_reverse
from awx.conf.license import get_license, feature_enabled
from awx.api.versioning import reverse, drf_reverse
from awx.conf.license import get_license
from awx.main.constants import PRIVILEGE_ESCALATION_METHODS
from awx.main.models import (
Project,
@@ -42,7 +42,7 @@ logger = logging.getLogger('awx.api.views.root')
class ApiRootView(APIView):
permission_classes = (AllowAny,)
view_name = _('REST API')
name = _('REST API')
versioning_class = None
swagger_topic = 'Versioning'
@@ -50,23 +50,21 @@ class ApiRootView(APIView):
def get(self, request, format=None):
''' List supported API versions '''
v1 = reverse('api:api_v1_root_view', kwargs={'version': 'v1'})
v2 = reverse('api:api_v2_root_view', kwargs={'version': 'v2'})
data = OrderedDict()
data['description'] = _('AWX REST API')
data['current_version'] = v2
data['available_versions'] = dict(v1 = v1, v2 = v2)
data['available_versions'] = dict(v2 = v2)
data['oauth2'] = drf_reverse('api:oauth_authorization_root_view')
if feature_enabled('rebranding'):
data['custom_logo'] = settings.CUSTOM_LOGO
data['custom_login_info'] = settings.CUSTOM_LOGIN_INFO
data['custom_logo'] = settings.CUSTOM_LOGO
data['custom_login_info'] = settings.CUSTOM_LOGIN_INFO
return Response(data)
class ApiOAuthAuthorizationRootView(APIView):
permission_classes = (AllowAny,)
view_name = _("API OAuth 2 Authorization Root")
name = _("API OAuth 2 Authorization Root")
versioning_class = None
swagger_topic = 'Authentication'
@@ -86,10 +84,10 @@ class ApiVersionRootView(APIView):
def get(self, request, format=None):
''' List top level resources '''
data = OrderedDict()
data['ping'] = reverse('api:api_v1_ping_view', request=request)
data['ping'] = reverse('api:api_v2_ping_view', request=request)
data['instances'] = reverse('api:instance_list', request=request)
data['instance_groups'] = reverse('api:instance_group_list', request=request)
data['config'] = reverse('api:api_v1_config_view', request=request)
data['config'] = reverse('api:api_v2_config_view', request=request)
data['settings'] = reverse('api:setting_category_list', request=request)
data['me'] = reverse('api:user_me_list', request=request)
data['dashboard'] = reverse('api:dashboard_view', request=request)
@@ -99,10 +97,11 @@ class ApiVersionRootView(APIView):
data['project_updates'] = reverse('api:project_update_list', request=request)
data['teams'] = reverse('api:team_list', request=request)
data['credentials'] = reverse('api:credential_list', request=request)
if get_request_version(request) > 1:
data['credential_types'] = reverse('api:credential_type_list', request=request)
data['applications'] = reverse('api:o_auth2_application_list', request=request)
data['tokens'] = reverse('api:o_auth2_token_list', request=request)
data['credential_types'] = reverse('api:credential_type_list', request=request)
data['credential_input_sources'] = reverse('api:credential_input_source_list', request=request)
data['applications'] = reverse('api:o_auth2_application_list', request=request)
data['tokens'] = reverse('api:o_auth2_token_list', request=request)
data['metrics'] = reverse('api:metrics_view', request=request)
data['inventory'] = reverse('api:inventory_list', request=request)
data['inventory_scripts'] = reverse('api:inventory_script_list', request=request)
data['inventory_sources'] = reverse('api:inventory_source_list', request=request)
@@ -130,21 +129,17 @@ class ApiVersionRootView(APIView):
return Response(data)
class ApiV1RootView(ApiVersionRootView):
view_name = _('Version 1')
class ApiV2RootView(ApiVersionRootView):
view_name = _('Version 2')
name = _('Version 2')
class ApiV1PingView(APIView):
class ApiV2PingView(APIView):
"""A simple view that reports very basic information about this
instance, which is acceptable to be public information.
"""
permission_classes = (AllowAny,)
authentication_classes = ()
view_name = _('Ping')
name = _('Ping')
swagger_topic = 'System Configuration'
def get(self, request, format=None):
@@ -157,29 +152,30 @@ class ApiV1PingView(APIView):
'ha': is_ha_environment(),
'version': get_awx_version(),
'active_node': settings.CLUSTER_HOST_ID,
'install_uuid': settings.INSTALL_UUID,
}
response['instances'] = []
for instance in Instance.objects.all():
response['instances'].append(dict(node=instance.hostname, heartbeat=instance.modified,
response['instances'].append(dict(node=instance.hostname, uuid=instance.uuid, heartbeat=instance.modified,
capacity=instance.capacity, version=instance.version))
sorted(response['instances'], key=operator.itemgetter('node'))
response['instance_groups'] = []
for instance_group in InstanceGroup.objects.all():
for instance_group in InstanceGroup.objects.prefetch_related('instances'):
response['instance_groups'].append(dict(name=instance_group.name,
capacity=instance_group.capacity,
instances=[x.hostname for x in instance_group.instances.all()]))
return Response(response)
class ApiV1ConfigView(APIView):
class ApiV2ConfigView(APIView):
permission_classes = (IsAuthenticated,)
view_name = _('Configuration')
name = _('Configuration')
swagger_topic = 'System Configuration'
def check_permissions(self, request):
super(ApiV1ConfigView, self).check_permissions(request)
super(ApiV2ConfigView, self).check_permissions(request)
if not request.user.is_superuser and request.method.lower() not in {'options', 'head', 'get'}:
self.permission_denied(request) # Raises PermissionDenied exception.
@@ -211,7 +207,7 @@ class ApiV1ConfigView(APIView):
# If LDAP is enabled, user_ldap_fields will return a list of field
# names that are managed by LDAP and should be read-only for users with
# a non-empty ldap_dn attribute.
if getattr(settings, 'AUTH_LDAP_SERVER_URI', None) and feature_enabled('ldap'):
if getattr(settings, 'AUTH_LDAP_SERVER_URI', None):
user_ldap_fields = ['username', 'password']
user_ldap_fields.extend(getattr(settings, 'AUTH_LDAP_USER_ATTR_MAP', {}).keys())
user_ldap_fields.extend(getattr(settings, 'AUTH_LDAP_USER_FLAGS_BY_GROUP', {}).keys())
@@ -220,7 +216,8 @@ class ApiV1ConfigView(APIView):
if request.user.is_superuser \
or request.user.is_system_auditor \
or Organization.accessible_objects(request.user, 'admin_role').exists() \
or Organization.accessible_objects(request.user, 'auditor_role').exists():
or Organization.accessible_objects(request.user, 'auditor_role').exists() \
or Organization.accessible_objects(request.user, 'project_admin_role').exists():
data.update(dict(
project_base_dir = settings.PROJECTS_ROOT,
project_local_paths = Project.get_local_path_choices(),
@@ -276,6 +273,3 @@ class ApiV1ConfigView(APIView):
except Exception:
# FIX: Log
return Response({"error": _("Failed to remove license.")}, status=status.HTTP_400_BAD_REQUEST)

View File

@@ -78,9 +78,6 @@ register(
# the other settings change, the cached value for this setting will be
# cleared to require it to be recomputed.
depends_on=['ANSIBLE_COW_SELECTION'],
# Optional; licensed feature required to be able to view or modify this
# setting.
feature_required='rebranding',
# Optional; field is stored encrypted in the database and only $encrypted$
# is returned via the API.
encrypted=True,

View File

@@ -1,4 +1,5 @@
# Python
import os
import logging
import urllib.parse as urlparse
from collections import OrderedDict
@@ -8,7 +9,10 @@ from django.core.validators import URLValidator
from django.utils.translation import ugettext_lazy as _
# Django REST Framework
from rest_framework.fields import * # noqa
from rest_framework.fields import ( # noqa
BooleanField, CharField, ChoiceField, DictField, EmailField, IntegerField,
ListField, NullBooleanField
)
logger = logging.getLogger('awx.conf.fields')
@@ -93,6 +97,26 @@ class StringListBooleanField(ListField):
self.fail('type_error', input_type=type(data))
class StringListPathField(StringListField):
default_error_messages = {
'type_error': _('Expected list of strings but got {input_type} instead.'),
'path_error': _('{path} is not a valid path choice.'),
}
def to_internal_value(self, paths):
if isinstance(paths, (list, tuple)):
for p in paths:
if not isinstance(p, str):
self.fail('type_error', input_type=type(p))
if not os.path.exists(p):
self.fail('path_error', path=p)
return super(StringListPathField, self).to_internal_value(sorted({os.path.normpath(path) for path in paths}))
else:
self.fail('type_error', input_type=type(paths))
class URLField(CharField):
def __init__(self, **kwargs):

View File

@@ -1,64 +1,19 @@
# Copyright (c) 2016 Ansible, Inc.
# All Rights Reserved.
# Django
from django.core.signals import setting_changed
from django.dispatch import receiver
from django.utils.translation import ugettext_lazy as _
# Django REST Framework
from rest_framework.exceptions import APIException
# Tower
from awx.main.utils.common import get_licenser
from awx.main.utils import memoize, memoize_delete
__all__ = ['LicenseForbids', 'get_license', 'get_licensed_features',
'feature_enabled', 'feature_exists']
class LicenseForbids(APIException):
status_code = 402
default_detail = _('Your Tower license does not allow that.')
__all__ = ['get_license']
def _get_validated_license_data():
return get_licenser().validate()
@receiver(setting_changed)
def _on_setting_changed(sender, **kwargs):
# Clear cached result above when license changes.
if kwargs.get('setting', None) == 'LICENSE':
memoize_delete('feature_enabled')
def get_license(show_key=False):
"""Return a dictionary representing the active license on this Tower instance."""
license_data = _get_validated_license_data()
if not show_key:
license_data.pop('license_key', None)
return license_data
def get_licensed_features():
"""Return a set of all features enabled by the active license."""
features = set()
for feature, enabled in _get_validated_license_data().get('features', {}).items():
if enabled:
features.add(feature)
return features
@memoize(track_function=True)
def feature_enabled(name):
"""Return True if the requested feature is enabled, False otherwise."""
validated_license_data = _get_validated_license_data()
if validated_license_data.get('license_type', 'UNLICENSED') == 'open':
return True
return validated_license_data.get('features', {}).get(name, False)
def feature_exists(name):
"""Return True if the requested feature name exists, False otherwise."""
return bool(name in _get_validated_license_data().get('features', {}))

View File

@@ -22,7 +22,7 @@ def fill_ldap_group_type_params(apps, schema_editor):
modified=now())
init_attrs = set(inspect.getargspec(group_type.__init__).args[1:])
for k in group_type_params.keys():
for k in list(group_type_params.keys()):
if k not in init_attrs:
del group_type_params[k]

View File

@@ -68,7 +68,7 @@ class SettingsRegistry(object):
def get_dependent_settings(self, setting):
return self._dependent_settings.get(setting, set())
def get_registered_categories(self, features_enabled=None):
def get_registered_categories(self):
categories = {
'all': _('All'),
'changed': _('Changed'),
@@ -77,10 +77,6 @@ class SettingsRegistry(object):
category_slug = kwargs.get('category_slug', None)
if category_slug is None or category_slug in categories:
continue
if features_enabled is not None:
feature_required = kwargs.get('feature_required', None)
if feature_required and feature_required not in features_enabled:
continue
if category_slug == 'user':
categories['user'] = _('User')
categories['user-defaults'] = _('User-Defaults')
@@ -88,7 +84,7 @@ class SettingsRegistry(object):
categories[category_slug] = kwargs.get('category', None) or category_slug
return categories
def get_registered_settings(self, category_slug=None, read_only=None, features_enabled=None, slugs_to_ignore=set()):
def get_registered_settings(self, category_slug=None, read_only=None, slugs_to_ignore=set()):
setting_names = []
if category_slug == 'user-defaults':
category_slug = 'user'
@@ -100,14 +96,10 @@ class SettingsRegistry(object):
if kwargs.get('category_slug', None) in slugs_to_ignore:
continue
if (read_only in {True, False} and kwargs.get('read_only', False) != read_only and
setting not in ('AWX_ISOLATED_PRIVATE_KEY', 'AWX_ISOLATED_PUBLIC_KEY')):
setting not in ('INSTALL_UUID', 'AWX_ISOLATED_PRIVATE_KEY', 'AWX_ISOLATED_PUBLIC_KEY')):
# Note: Doesn't catch fields that set read_only via __init__;
# read-only field kwargs should always include read_only=True.
continue
if features_enabled is not None:
feature_required = kwargs.get('feature_required', None)
if feature_required and feature_required not in features_enabled:
continue
setting_names.append(setting)
return setting_names
@@ -135,7 +127,6 @@ class SettingsRegistry(object):
category = field_kwargs.pop('category', None)
depends_on = frozenset(field_kwargs.pop('depends_on', None) or [])
placeholder = field_kwargs.pop('placeholder', empty)
feature_required = field_kwargs.pop('feature_required', empty)
encrypted = bool(field_kwargs.pop('encrypted', False))
defined_in_file = bool(field_kwargs.pop('defined_in_file', False))
if getattr(field_kwargs.get('child', None), 'source', None) is not None:
@@ -146,8 +137,6 @@ class SettingsRegistry(object):
field_instance.depends_on = depends_on
if placeholder is not empty:
field_instance.placeholder = placeholder
if feature_required is not empty:
field_instance.feature_required = feature_required
field_instance.defined_in_file = defined_in_file
if field_instance.defined_in_file:
field_instance.help_text = (

View File

@@ -88,7 +88,7 @@ class SettingSingletonSerializer(serializers.Serializer):
continue
extra_kwargs = {}
# Make LICENSE and AWX_ISOLATED_KEY_GENERATION read-only here;
# LICENSE is only updated via /api/v1/config/
# LICENSE is only updated via /api/v2/config/
# AWX_ISOLATED_KEY_GENERATION is only set/unset via the setup playbook
if key in ('LICENSE', 'AWX_ISOLATED_KEY_GENERATION'):
extra_kwargs['read_only'] = True

View File

@@ -24,7 +24,6 @@ from rest_framework.fields import empty, SkipField
# Tower
from awx.main.utils import encrypt_field, decrypt_field
from awx.main.utils.db import get_tower_migration_version
from awx.conf import settings_registry
from awx.conf.models import Setting
from awx.conf.migrations._reencrypt import decrypt_field as old_decrypt_field
@@ -90,45 +89,42 @@ def _ctit_db_wrapper(trans_safe=False):
transaction.set_rollback(False)
yield
except DBError:
if 'migrate' in sys.argv and get_tower_migration_version() < '310':
logger.info('Using default settings until version 3.1 migration.')
else:
# We want the _full_ traceback with the context
# First we get the current call stack, which constitutes the "top",
# it has the context up to the point where the context manager is used
top_stack = StringIO()
traceback.print_stack(file=top_stack)
top_lines = top_stack.getvalue().strip('\n').split('\n')
top_stack.close()
# Get "bottom" stack from the local error that happened
# inside of the "with" block this wraps
exc_type, exc_value, exc_traceback = sys.exc_info()
bottom_stack = StringIO()
traceback.print_tb(exc_traceback, file=bottom_stack)
bottom_lines = bottom_stack.getvalue().strip('\n').split('\n')
# Glue together top and bottom where overlap is found
bottom_cutoff = 0
for i, line in enumerate(bottom_lines):
if line in top_lines:
# start of overlapping section, take overlap from bottom
top_lines = top_lines[:top_lines.index(line)]
bottom_cutoff = i
break
bottom_lines = bottom_lines[bottom_cutoff:]
tb_lines = top_lines + bottom_lines
# We want the _full_ traceback with the context
# First we get the current call stack, which constitutes the "top",
# it has the context up to the point where the context manager is used
top_stack = StringIO()
traceback.print_stack(file=top_stack)
top_lines = top_stack.getvalue().strip('\n').split('\n')
top_stack.close()
# Get "bottom" stack from the local error that happened
# inside of the "with" block this wraps
exc_type, exc_value, exc_traceback = sys.exc_info()
bottom_stack = StringIO()
traceback.print_tb(exc_traceback, file=bottom_stack)
bottom_lines = bottom_stack.getvalue().strip('\n').split('\n')
# Glue together top and bottom where overlap is found
bottom_cutoff = 0
for i, line in enumerate(bottom_lines):
if line in top_lines:
# start of overlapping section, take overlap from bottom
top_lines = top_lines[:top_lines.index(line)]
bottom_cutoff = i
break
bottom_lines = bottom_lines[bottom_cutoff:]
tb_lines = top_lines + bottom_lines
tb_string = '\n'.join(
['Traceback (most recent call last):'] +
tb_lines +
['{}: {}'.format(exc_type.__name__, str(exc_value))]
)
bottom_stack.close()
# Log the combined stack
if trans_safe:
if 'check_migrations' not in sys.argv:
logger.warning('Database settings are not available, using defaults, error:\n{}'.format(tb_string))
else:
logger.error('Error modifying something related to database settings.\n{}'.format(tb_string))
tb_string = '\n'.join(
['Traceback (most recent call last):'] +
tb_lines +
['{}: {}'.format(exc_type.__name__, str(exc_value))]
)
bottom_stack.close()
# Log the combined stack
if trans_safe:
if 'check_migrations' not in sys.argv:
logger.debug('Database settings are not available, using defaults, error:\n{}'.format(tb_string))
else:
logger.debug('Error modifying something related to database settings.\n{}'.format(tb_string))
finally:
if trans_safe and is_atomic and rollback_set:
transaction.set_rollback(rollback_set)
@@ -381,8 +377,9 @@ class SettingsWrapper(UserSettingsHolder):
setting = None
setting_id = None
if not field.read_only or name in (
# these two values are read-only - however - we *do* want
# these values are read-only - however - we *do* want
# to fetch their value from the database
'INSTALL_UUID',
'AWX_ISOLATED_PRIVATE_KEY',
'AWX_ISOLATED_PUBLIC_KEY',
):

View File

@@ -2,7 +2,7 @@ import urllib.parse
import pytest
from django.core.urlresolvers import resolve
from django.urls import resolve
from django.contrib.auth.models import User
from rest_framework.test import (

View File

@@ -65,41 +65,6 @@ def test_non_admin_user_does_not_see_categories(api_request, dummy_setting, norm
assert not response.data['results']
@pytest.mark.django_db
@mock.patch(
'awx.conf.views.VERSION_SPECIFIC_CATEGORIES_TO_EXCLUDE',
{
1: set([]),
2: set(['foobar']),
}
)
def test_version_specific_category_slug_to_exclude_does_not_show_up(api_request, dummy_setting):
with dummy_setting(
'FOO_BAR',
field_class=fields.IntegerField,
category='FooBar',
category_slug='foobar'
):
response = api_request(
'get',
reverse('api:setting_category_list',
kwargs={'version': 'v2'})
)
for item in response.data['results']:
assert item['slug'] != 'foobar'
response = api_request(
'get',
reverse('api:setting_category_list',
kwargs={'version': 'v1'})
)
contains = False
for item in response.data['results']:
if item['slug'] != 'foobar':
contains = True
break
assert contains
@pytest.mark.django_db
def test_setting_singleton_detail_retrieve(api_request, dummy_setting):
with dummy_setting(

View File

@@ -1,7 +1,7 @@
import pytest
from rest_framework.fields import ValidationError
from awx.conf.fields import StringListBooleanField, ListTuplesField
from awx.conf.fields import StringListBooleanField, StringListPathField, ListTuplesField
class TestStringListBooleanField():
@@ -84,3 +84,49 @@ class TestListTuplesField():
assert e.value.detail[0] == "Expected a list of tuples of max length 2 " \
"but got {} instead.".format(t)
class TestStringListPathField():
FIELD_VALUES = [
((".", "..", "/"), [".", "..", "/"]),
(("/home",), ["/home"]),
(("///home///",), ["/home"]),
(("/home/././././",), ["/home"]),
(("/home", "/home", "/home/"), ["/home"]),
(["/home/", "/home/", "/opt/", "/opt/", "/var/"], ["/home", "/opt", "/var"])
]
FIELD_VALUES_INVALID_TYPE = [
1.245,
{"a": "b"},
("/home"),
]
FIELD_VALUES_INVALID_PATH = [
"",
"~/",
"home",
"/invalid_path",
"/home/invalid_path",
]
@pytest.mark.parametrize("value_in, value_known", FIELD_VALUES)
def test_to_internal_value_valid(self, value_in, value_known):
field = StringListPathField()
v = field.to_internal_value(value_in)
assert v == value_known
@pytest.mark.parametrize("value", FIELD_VALUES_INVALID_TYPE)
def test_to_internal_value_invalid_type(self, value):
field = StringListPathField()
with pytest.raises(ValidationError) as e:
field.to_internal_value(value)
assert e.value.detail[0] == "Expected list of strings but got {} instead.".format(type(value))
@pytest.mark.parametrize("value", FIELD_VALUES_INVALID_PATH)
def test_to_internal_value_invalid_path(self, value):
field = StringListPathField()
with pytest.raises(ValidationError) as e:
field.to_internal_value([value])
assert e.value.detail[0] == "{} is not a valid path choice.".format(value)

View File

@@ -119,20 +119,6 @@ def test_get_registered_read_only_settings(reg):
]
def test_get_registered_settings_with_required_features(reg):
reg.register(
'AWX_SOME_SETTING_ENABLED',
field_class=fields.BooleanField,
category=_('System'),
category_slug='system',
feature_required='superpowers',
)
assert reg.get_registered_settings(features_enabled=[]) == []
assert reg.get_registered_settings(features_enabled=['superpowers']) == [
'AWX_SOME_SETTING_ENABLED'
]
def test_get_dependent_settings(reg):
reg.register(
'AWX_SOME_SETTING_ENABLED',
@@ -173,45 +159,6 @@ def test_get_registered_categories(reg):
}
def test_get_registered_categories_with_required_features(reg):
reg.register(
'AWX_SOME_SETTING_ENABLED',
field_class=fields.BooleanField,
category=_('System'),
category_slug='system',
feature_required='superpowers'
)
reg.register(
'AWX_SOME_OTHER_SETTING_ENABLED',
field_class=fields.BooleanField,
category=_('OtherSystem'),
category_slug='other-system',
feature_required='sortapowers'
)
assert reg.get_registered_categories(features_enabled=[]) == {
'all': _('All'),
'changed': _('Changed'),
}
assert reg.get_registered_categories(features_enabled=['superpowers']) == {
'all': _('All'),
'changed': _('Changed'),
'system': _('System'),
}
assert reg.get_registered_categories(features_enabled=['sortapowers']) == {
'all': _('All'),
'changed': _('Changed'),
'other-system': _('OtherSystem'),
}
assert reg.get_registered_categories(
features_enabled=['superpowers', 'sortapowers']
) == {
'all': _('All'),
'changed': _('Changed'),
'system': _('System'),
'other-system': _('OtherSystem'),
}
def test_is_setting_encrypted(reg):
reg.register(
'AWX_SOME_SETTING_ENABLED',
@@ -237,7 +184,6 @@ def test_simple_field(reg):
category=_('System'),
category_slug='system',
placeholder='Example Value',
feature_required='superpowers'
)
field = reg.get_setting_field('AWX_SOME_SETTING')
@@ -246,7 +192,6 @@ def test_simple_field(reg):
assert field.category_slug == 'system'
assert field.default is empty
assert field.placeholder == 'Example Value'
assert field.feature_required == 'superpowers'
def test_field_with_custom_attribute(reg):

View File

@@ -1,108 +1,9 @@
#!/usr/bin/env python
# Python
import difflib
import glob
import os
import shutil
# AWX
from awx.conf.registry import settings_registry
__all__ = ['comment_assignments', 'conf_to_dict']
def comment_assignments(patterns, assignment_names, dry_run=True, backup_suffix='.old'):
if isinstance(patterns, str):
patterns = [patterns]
diffs = []
for pattern in patterns:
for filename in sorted(glob.glob(pattern)):
filename = os.path.abspath(os.path.normpath(filename))
if backup_suffix:
backup_filename = '{}{}'.format(filename, backup_suffix)
else:
backup_filename = None
diff = comment_assignments_in_file(filename, assignment_names, dry_run, backup_filename)
if diff:
diffs.append(diff)
return diffs
def comment_assignments_in_file(filename, assignment_names, dry_run=True, backup_filename=None):
from redbaron import RedBaron, indent
if isinstance(assignment_names, str):
assignment_names = [assignment_names]
else:
assignment_names = assignment_names[:]
current_file_data = open(filename).read()
for assignment_name in assignment_names[:]:
if assignment_name in current_file_data:
continue
if assignment_name in assignment_names:
assignment_names.remove(assignment_name)
if not assignment_names:
return ''
replace_lines = {}
rb = RedBaron(current_file_data)
for assignment_node in rb.find_all('assignment'):
for assignment_name in assignment_names:
# Only target direct assignments to a variable.
name_node = assignment_node.find('name', value=assignment_name)
if not name_node:
continue
if assignment_node.target.type != 'name':
continue
# Build a new node that comments out the existing assignment node.
indentation = '{}# '.format(assignment_node.indentation or '')
new_node_content = indent(assignment_node.dumps(), indentation)
new_node_lines = new_node_content.splitlines()
# Add a pass statement in case the assignment block is the only
# child in a parent code block to prevent a syntax error.
if assignment_node.indentation:
new_node_lines[0] = new_node_lines[0].replace(indentation, '{}pass # '.format(assignment_node.indentation or ''), 1)
new_node_lines[0] = '{0}This setting is now configured via the Tower API.\n{1}'.format(indentation, new_node_lines[0])
# Store new node lines in dictionary to be replaced in file.
start_lineno = assignment_node.absolute_bounding_box.top_left.line
end_lineno = assignment_node.absolute_bounding_box.bottom_right.line
for n, new_node_line in enumerate(new_node_lines):
new_lineno = start_lineno + n
assert new_lineno <= end_lineno
replace_lines[new_lineno] = new_node_line
if not replace_lines:
return ''
# Iterate through all lines in current file and replace as needed.
current_file_lines = current_file_data.splitlines()
new_file_lines = []
for n, line in enumerate(current_file_lines):
new_file_lines.append(replace_lines.get(n + 1, line))
new_file_data = '\n'.join(new_file_lines)
new_file_lines = new_file_data.splitlines()
# If changed, syntax check and write the new file; return a diff of changes.
diff_lines = []
if new_file_data != current_file_data:
compile(new_file_data, filename, 'exec')
if backup_filename:
from_file = backup_filename
else:
from_file = '{}.old'.format(filename)
to_file = filename
diff_lines = list(difflib.unified_diff(current_file_lines, new_file_lines, fromfile=from_file, tofile=to_file, lineterm=''))
if not dry_run:
if backup_filename:
shutil.copy2(filename, backup_filename)
with open(filename, 'w') as fileobj:
fileobj.write(new_file_data)
return '\n'.join(diff_lines)
__all__ = ['conf_to_dict']
def conf_to_dict(obj):
@@ -110,10 +11,3 @@ def conf_to_dict(obj):
'category': settings_registry.get_setting_category(obj.key),
'name': obj.key,
}
if __name__ == '__main__':
pattern = os.path.join(os.path.dirname(__file__), '..', 'settings', 'local_*.py')
diffs = comment_assignments(pattern, ['AUTH_LDAP_ORGANIZATION_MAP'])
for diff in diffs:
print(diff)

View File

@@ -17,13 +17,17 @@ from rest_framework import serializers
from rest_framework import status
# Tower
from awx.api.generics import * # noqa
from awx.api.generics import (
APIView,
GenericAPIView,
ListAPIView,
RetrieveUpdateDestroyAPIView,
)
from awx.api.permissions import IsSuperUser
from awx.api.versioning import reverse, get_request_version
from awx.main.utils import * # noqa
from awx.api.versioning import reverse
from awx.main.utils import camelcase_to_underscore
from awx.main.utils.handlers import AWXProxyHandler, LoggingConnectivityException
from awx.main.tasks import handle_setting_changes
from awx.conf.license import get_licensed_features
from awx.conf.models import Setting
from awx.conf.serializers import SettingCategorySerializer, SettingSingletonSerializer
from awx.conf import settings_registry
@@ -31,24 +35,17 @@ from awx.conf import settings_registry
SettingCategory = collections.namedtuple('SettingCategory', ('url', 'slug', 'name'))
VERSION_SPECIFIC_CATEGORIES_TO_EXCLUDE = {
1: set([
'named-url',
]),
2: set([]),
}
class SettingCategoryList(ListAPIView):
model = Setting # Not exactly, but needed for the view.
serializer_class = SettingCategorySerializer
filter_backends = []
view_name = _('Setting Categories')
name = _('Setting Categories')
def get_queryset(self):
setting_categories = []
categories = settings_registry.get_registered_categories(features_enabled=get_licensed_features())
categories = settings_registry.get_registered_categories()
if self.request.user.is_superuser or self.request.user.is_system_auditor:
pass # categories = categories
elif 'user' in categories:
@@ -56,8 +53,6 @@ class SettingCategoryList(ListAPIView):
else:
categories = {}
for category_slug in sorted(categories.keys()):
if category_slug in VERSION_SPECIFIC_CATEGORIES_TO_EXCLUDE[get_request_version(self.request)]:
continue
url = reverse('api:setting_singleton_detail', kwargs={'category_slug': category_slug}, request=self.request)
setting_categories.append(SettingCategory(url, category_slug, categories[category_slug]))
return setting_categories
@@ -68,13 +63,11 @@ class SettingSingletonDetail(RetrieveUpdateDestroyAPIView):
model = Setting # Not exactly, but needed for the view.
serializer_class = SettingSingletonSerializer
filter_backends = []
view_name = _('Setting Detail')
name = _('Setting Detail')
def get_queryset(self):
self.category_slug = self.kwargs.get('category_slug', 'all')
all_category_slugs = list(settings_registry.get_registered_categories(features_enabled=get_licensed_features()).keys())
for slug_to_delete in VERSION_SPECIFIC_CATEGORIES_TO_EXCLUDE[get_request_version(self.request)]:
all_category_slugs.remove(slug_to_delete)
all_category_slugs = list(settings_registry.get_registered_categories().keys())
if self.request.user.is_superuser or getattr(self.request.user, 'is_system_auditor', False):
category_slugs = all_category_slugs
else:
@@ -85,8 +78,7 @@ class SettingSingletonDetail(RetrieveUpdateDestroyAPIView):
raise PermissionDenied()
registered_settings = settings_registry.get_registered_settings(
category_slug=self.category_slug, read_only=False, features_enabled=get_licensed_features(),
slugs_to_ignore=VERSION_SPECIFIC_CATEGORIES_TO_EXCLUDE[get_request_version(self.request)]
category_slug=self.category_slug, read_only=False,
)
if self.category_slug == 'user':
return Setting.objects.filter(key__in=registered_settings, user=self.request.user)
@@ -96,8 +88,7 @@ class SettingSingletonDetail(RetrieveUpdateDestroyAPIView):
def get_object(self):
settings_qs = self.get_queryset()
registered_settings = settings_registry.get_registered_settings(
category_slug=self.category_slug, features_enabled=get_licensed_features(),
slugs_to_ignore=VERSION_SPECIFIC_CATEGORIES_TO_EXCLUDE[get_request_version(self.request)]
category_slug=self.category_slug,
)
all_settings = {}
for setting in settings_qs:
@@ -163,7 +154,7 @@ class SettingSingletonDetail(RetrieveUpdateDestroyAPIView):
class SettingLoggingTest(GenericAPIView):
view_name = _('Logging Connectivity Test')
name = _('Logging Connectivity Test')
model = Setting
serializer_class = SettingSingletonSerializer
permission_classes = (IsSuperUser,)

View File

@@ -1,25 +0,0 @@
# Copyright (c) 2016 Ansible by Red Hat, Inc.
#
# This file is part of Ansible Tower, but depends on code imported from Ansible.
#
# Ansible is free software: you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# Ansible is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with Ansible. If not, see <http://www.gnu.org/licenses/>.
from __future__ import (absolute_import, division, print_function)
# AWX Display Callback
from . import cleanup # noqa (registers control persistent cleanup)
from . import display # noqa (wraps ansible.display.Display methods)
from .module import AWXDefaultCallbackModule, AWXMinimalCallbackModule
__all__ = ['AWXDefaultCallbackModule', 'AWXMinimalCallbackModule']

View File

@@ -1,85 +0,0 @@
# Copyright (c) 2016 Ansible by Red Hat, Inc.
#
# This file is part of Ansible Tower, but depends on code imported from Ansible.
#
# Ansible is free software: you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# Ansible is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with Ansible. If not, see <http://www.gnu.org/licenses/>.
from __future__ import (absolute_import, division, print_function)
# Python
import atexit
import glob
import os
import pwd
# PSUtil
try:
import psutil
except ImportError:
raise ImportError('psutil is missing; {}bin/pip install psutil'.format(
os.environ['VIRTUAL_ENV']
))
__all__ = []
main_pid = os.getpid()
@atexit.register
def terminate_ssh_control_masters():
# Only run this cleanup from the main process.
if os.getpid() != main_pid:
return
# Determine if control persist is being used and if any open sockets
# exist after running the playbook.
cp_path = os.environ.get('ANSIBLE_SSH_CONTROL_PATH', '')
if not cp_path:
return
cp_dir = os.path.dirname(cp_path)
if not os.path.exists(cp_dir):
return
cp_pattern = os.path.join(cp_dir, 'ansible-ssh-*')
cp_files = glob.glob(cp_pattern)
if not cp_files:
return
# Attempt to find any running control master processes.
username = pwd.getpwuid(os.getuid())[0]
ssh_cm_procs = []
for proc in psutil.process_iter():
try:
pname = proc.name()
pcmdline = proc.cmdline()
pusername = proc.username()
except psutil.NoSuchProcess:
continue
if pusername != username:
continue
if pname != 'ssh':
continue
for cp_file in cp_files:
if pcmdline and cp_file in pcmdline[0]:
ssh_cm_procs.append(proc)
break
# Terminate then kill control master processes. Workaround older
# version of psutil that may not have wait_procs implemented.
for proc in ssh_cm_procs:
try:
proc.terminate()
except psutil.NoSuchProcess:
continue
procs_gone, procs_alive = psutil.wait_procs(ssh_cm_procs, timeout=5)
for proc in procs_alive:
proc.kill()

View File

@@ -1,98 +0,0 @@
# Copyright (c) 2016 Ansible by Red Hat, Inc.
#
# This file is part of Ansible Tower, but depends on code imported from Ansible.
#
# Ansible is free software: you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# Ansible is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with Ansible. If not, see <http://www.gnu.org/licenses/>.
from __future__ import (absolute_import, division, print_function)
# Python
import functools
import sys
import uuid
# Ansible
from ansible.utils.display import Display
# Tower Display Callback
from .events import event_context
__all__ = []
def with_context(**context):
global event_context
def wrap(f):
@functools.wraps(f)
def wrapper(*args, **kwargs):
with event_context.set_local(**context):
return f(*args, **kwargs)
return wrapper
return wrap
for attr in dir(Display):
if attr.startswith('_') or 'cow' in attr or 'prompt' in attr:
continue
if attr in ('display', 'v', 'vv', 'vvv', 'vvvv', 'vvvvv', 'vvvvvv', 'verbose'):
continue
if not callable(getattr(Display, attr)):
continue
setattr(Display, attr, with_context(**{attr: True})(getattr(Display, attr)))
def with_verbosity(f):
global event_context
@functools.wraps(f)
def wrapper(*args, **kwargs):
host = args[2] if len(args) >= 3 else kwargs.get('host', None)
caplevel = args[3] if len(args) >= 4 else kwargs.get('caplevel', 2)
context = dict(verbose=True, verbosity=(caplevel + 1))
if host is not None:
context['remote_addr'] = host
with event_context.set_local(**context):
return f(*args, **kwargs)
return wrapper
Display.verbose = with_verbosity(Display.verbose)
def display_with_context(f):
@functools.wraps(f)
def wrapper(*args, **kwargs):
log_only = args[5] if len(args) >= 6 else kwargs.get('log_only', False)
stderr = args[3] if len(args) >= 4 else kwargs.get('stderr', False)
event_uuid = event_context.get().get('uuid', None)
with event_context.display_lock:
# If writing only to a log file or there is already an event UUID
# set (from a callback module method), skip dumping the event data.
if log_only or event_uuid:
return f(*args, **kwargs)
try:
fileobj = sys.stderr if stderr else sys.stdout
event_context.add_local(uuid=str(uuid.uuid4()))
event_context.dump_begin(fileobj)
return f(*args, **kwargs)
finally:
event_context.dump_end(fileobj)
event_context.remove_local(uuid=None)
return wrapper
Display.display = display_with_context(Display.display)

View File

@@ -1,186 +0,0 @@
# Copyright (c) 2016 Ansible by Red Hat, Inc.
#
# This file is part of Ansible Tower, but depends on code imported from Ansible.
#
# Ansible is free software: you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# Ansible is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with Ansible. If not, see <http://www.gnu.org/licenses/>.
from __future__ import (absolute_import, division, print_function)
# Python
import base64
import contextlib
import datetime
import json
import multiprocessing
import os
import stat
import threading
import uuid
try:
import memcache
except ImportError:
raise ImportError('python-memcached is missing; {}bin/pip install python-memcached'.format(
os.environ['VIRTUAL_ENV']
))
__all__ = ['event_context']
class IsolatedFileWrite:
'''
Stand-in class that will write partial event data to a file as a
replacement for memcache when a job is running on an isolated host.
'''
def __init__(self):
self.private_data_dir = os.getenv('AWX_ISOLATED_DATA_DIR')
def set(self, key, value):
# Strip off the leading memcache key identifying characters :1:ev-
event_uuid = key[len(':1:ev-'):]
# Write data in a staging area and then atomic move to pickup directory
filename = '{}-partial.json'.format(event_uuid)
dropoff_location = os.path.join(self.private_data_dir, 'artifacts', 'job_events', filename)
write_location = '.'.join([dropoff_location, 'tmp'])
with os.fdopen(os.open(write_location, os.O_WRONLY | os.O_CREAT, stat.S_IRUSR | stat.S_IWUSR), 'w') as f:
f.write(value)
os.rename(write_location, dropoff_location)
class EventContext(object):
'''
Store global and local (per thread/process) data associated with callback
events and other display output methods.
'''
def __init__(self):
self.display_lock = multiprocessing.RLock()
cache_actual = os.getenv('CACHE', '127.0.0.1:11211')
if os.getenv('AWX_ISOLATED_DATA_DIR', False):
self.cache = IsolatedFileWrite()
else:
self.cache = memcache.Client([cache_actual], debug=0)
def add_local(self, **kwargs):
if not hasattr(self, '_local'):
self._local = threading.local()
self._local._ctx = {}
self._local._ctx.update(kwargs)
def remove_local(self, **kwargs):
if hasattr(self, '_local'):
for key in kwargs.keys():
self._local._ctx.pop(key, None)
@contextlib.contextmanager
def set_local(self, **kwargs):
try:
self.add_local(**kwargs)
yield
finally:
self.remove_local(**kwargs)
def get_local(self):
return getattr(getattr(self, '_local', None), '_ctx', {})
def add_global(self, **kwargs):
if not hasattr(self, '_global_ctx'):
self._global_ctx = {}
self._global_ctx.update(kwargs)
def remove_global(self, **kwargs):
if hasattr(self, '_global_ctx'):
for key in kwargs.keys():
self._global_ctx.pop(key, None)
@contextlib.contextmanager
def set_global(self, **kwargs):
try:
self.add_global(**kwargs)
yield
finally:
self.remove_global(**kwargs)
def get_global(self):
return getattr(self, '_global_ctx', {})
def get(self):
ctx = {}
ctx.update(self.get_global())
ctx.update(self.get_local())
return ctx
def get_begin_dict(self):
event_data = self.get()
if os.getenv('JOB_ID', ''):
event_data['job_id'] = int(os.getenv('JOB_ID', '0'))
if os.getenv('AD_HOC_COMMAND_ID', ''):
event_data['ad_hoc_command_id'] = int(os.getenv('AD_HOC_COMMAND_ID', '0'))
if os.getenv('PROJECT_UPDATE_ID', ''):
event_data['project_update_id'] = int(os.getenv('PROJECT_UPDATE_ID', '0'))
event_data.setdefault('pid', os.getpid())
event_data.setdefault('uuid', str(uuid.uuid4()))
event_data.setdefault('created', datetime.datetime.utcnow().isoformat())
if not event_data.get('parent_uuid', None) and event_data.get('job_id', None):
for key in ('task_uuid', 'play_uuid', 'playbook_uuid'):
parent_uuid = event_data.get(key, None)
if parent_uuid and parent_uuid != event_data.get('uuid', None):
event_data['parent_uuid'] = parent_uuid
break
event = event_data.pop('event', None)
if not event:
event = 'verbose'
for key in ('debug', 'verbose', 'deprecated', 'warning', 'system_warning', 'error'):
if event_data.get(key, False):
event = key
break
max_res = int(os.getenv("MAX_EVENT_RES", 700000))
if event not in ('playbook_on_stats',) and "res" in event_data and len(str(event_data['res'])) > max_res:
event_data['res'] = {}
event_dict = dict(event=event, event_data=event_data)
for key in list(event_data.keys()):
if key in ('job_id', 'ad_hoc_command_id', 'project_update_id', 'uuid', 'parent_uuid', 'created',):
event_dict[key] = event_data.pop(key)
elif key in ('verbosity', 'pid'):
event_dict[key] = event_data[key]
return event_dict
def get_end_dict(self):
return {}
def dump(self, fileobj, data, max_width=78, flush=False):
b64data = base64.b64encode(json.dumps(data).encode('utf-8')).decode()
with self.display_lock:
# pattern corresponding to OutputEventFilter expectation
fileobj.write(u'\x1b[K')
for offset in range(0, len(b64data), max_width):
chunk = b64data[offset:offset + max_width]
escaped_chunk = u'{}\x1b[{}D'.format(chunk, len(chunk))
fileobj.write(escaped_chunk)
fileobj.write(u'\x1b[K')
if flush:
fileobj.flush()
def dump_begin(self, fileobj):
begin_dict = self.get_begin_dict()
self.cache.set(":1:ev-{}".format(begin_dict['uuid']), json.dumps(begin_dict))
self.dump(fileobj, {'uuid': begin_dict['uuid']})
def dump_end(self, fileobj):
self.dump(fileobj, self.get_end_dict(), flush=True)
event_context = EventContext()

View File

@@ -1,29 +0,0 @@
# Copyright (c) 2016 Ansible by Red Hat, Inc.
#
# This file is part of Ansible Tower, but depends on code imported from Ansible.
#
# Ansible is free software: you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# Ansible is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with Ansible. If not, see <http://www.gnu.org/licenses/>.
from __future__ import (absolute_import, division, print_function)
# Python
import os
# Ansible
import ansible
# Because of the way Ansible loads plugins, it's not possible to import
# ansible.plugins.callback.minimal when being loaded as the minimal plugin. Ugh.
with open(os.path.join(os.path.dirname(ansible.__file__), 'plugins', 'callback', 'minimal.py')) as in_file:
exec(in_file.read())

View File

@@ -1,501 +0,0 @@
# Copyright (c) 2016 Ansible by Red Hat, Inc.
#
# This file is part of Ansible Tower, but depends on code imported from Ansible.
#
# Ansible is free software: you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# Ansible is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with Ansible. If not, see <http://www.gnu.org/licenses/>.
from __future__ import (absolute_import, division, print_function)
# Python
import codecs
import contextlib
import json
import os
import stat
import sys
import uuid
from copy import copy
# Ansible
from ansible import constants as C
from ansible.plugins.callback import CallbackBase
from ansible.plugins.callback.default import CallbackModule as DefaultCallbackModule
# AWX Display Callback
from .events import event_context
from .minimal import CallbackModule as MinimalCallbackModule
CENSORED = "the output has been hidden due to the fact that 'no_log: true' was specified for this result" # noqa
class BaseCallbackModule(CallbackBase):
'''
Callback module for logging ansible/ansible-playbook events.
'''
CALLBACK_VERSION = 2.0
CALLBACK_TYPE = 'stdout'
# These events should never have an associated play.
EVENTS_WITHOUT_PLAY = [
'playbook_on_start',
'playbook_on_stats',
]
# These events should never have an associated task.
EVENTS_WITHOUT_TASK = EVENTS_WITHOUT_PLAY + [
'playbook_on_setup',
'playbook_on_notify',
'playbook_on_import_for_host',
'playbook_on_not_import_for_host',
'playbook_on_no_hosts_matched',
'playbook_on_no_hosts_remaining',
]
def __init__(self):
super(BaseCallbackModule, self).__init__()
self.task_uuids = set()
@contextlib.contextmanager
def capture_event_data(self, event, **event_data):
event_data.setdefault('uuid', str(uuid.uuid4()))
if event not in self.EVENTS_WITHOUT_TASK:
task = event_data.pop('task', None)
else:
task = None
if event_data.get('res'):
if event_data['res'].get('_ansible_no_log', False):
event_data['res'] = {'censored': CENSORED}
if event_data['res'].get('results', []):
event_data['res']['results'] = copy(event_data['res']['results'])
for i, item in enumerate(event_data['res'].get('results', [])):
if isinstance(item, dict) and item.get('_ansible_no_log', False):
event_data['res']['results'][i] = {'censored': CENSORED}
with event_context.display_lock:
try:
event_context.add_local(event=event, **event_data)
if task:
self.set_task(task, local=True)
event_context.dump_begin(sys.stdout)
yield
finally:
event_context.dump_end(sys.stdout)
if task:
self.clear_task(local=True)
event_context.remove_local(event=None, **event_data)
def set_playbook(self, playbook):
# NOTE: Ansible doesn't generate a UUID for playbook_on_start so do it for them.
self.playbook_uuid = str(uuid.uuid4())
file_name = getattr(playbook, '_file_name', '???')
event_context.add_global(playbook=file_name, playbook_uuid=self.playbook_uuid)
self.clear_play()
def set_play(self, play):
if hasattr(play, 'hosts'):
if isinstance(play.hosts, list):
pattern = ','.join(play.hosts)
else:
pattern = play.hosts
else:
pattern = ''
name = play.get_name().strip() or pattern
event_context.add_global(play=name, play_uuid=str(play._uuid), play_pattern=pattern)
self.clear_task()
def clear_play(self):
event_context.remove_global(play=None, play_uuid=None, play_pattern=None)
self.clear_task()
def set_task(self, task, local=False):
# FIXME: Task is "global" unless using free strategy!
task_ctx = dict(
task=(task.name or task.action),
task_uuid=str(task._uuid),
task_action=task.action,
task_args='',
)
try:
task_ctx['task_path'] = task.get_path()
except AttributeError:
pass
if C.DISPLAY_ARGS_TO_STDOUT:
if task.no_log:
task_ctx['task_args'] = "the output has been hidden due to the fact that 'no_log: true' was specified for this result"
else:
task_args = ', '.join(('%s=%s' % a for a in task.args.items()))
task_ctx['task_args'] = task_args
if getattr(task, '_role', None):
task_role = task._role._role_name
else:
task_role = getattr(task, 'role_name', '')
if task_role:
task_ctx['role'] = task_role
if local:
event_context.add_local(**task_ctx)
else:
event_context.add_global(**task_ctx)
def clear_task(self, local=False):
task_ctx = dict(task=None, task_path=None, task_uuid=None, task_action=None, task_args=None, role=None)
if local:
event_context.remove_local(**task_ctx)
else:
event_context.remove_global(**task_ctx)
def v2_playbook_on_start(self, playbook):
self.set_playbook(playbook)
event_data = dict(
uuid=self.playbook_uuid,
)
with self.capture_event_data('playbook_on_start', **event_data):
super(BaseCallbackModule, self).v2_playbook_on_start(playbook)
def v2_playbook_on_vars_prompt(self, varname, private=True, prompt=None,
encrypt=None, confirm=False, salt_size=None,
salt=None, default=None):
event_data = dict(
varname=varname,
private=private,
prompt=prompt,
encrypt=encrypt,
confirm=confirm,
salt_size=salt_size,
salt=salt,
default=default,
)
with self.capture_event_data('playbook_on_vars_prompt', **event_data):
super(BaseCallbackModule, self).v2_playbook_on_vars_prompt(
varname, private, prompt, encrypt, confirm, salt_size, salt,
default,
)
def v2_playbook_on_include(self, included_file):
event_data = dict(
included_file=included_file._filename if included_file is not None else None,
)
with self.capture_event_data('playbook_on_include', **event_data):
super(BaseCallbackModule, self).v2_playbook_on_include(included_file)
def v2_playbook_on_play_start(self, play):
self.set_play(play)
if hasattr(play, 'hosts'):
if isinstance(play.hosts, list):
pattern = ','.join(play.hosts)
else:
pattern = play.hosts
else:
pattern = ''
name = play.get_name().strip() or pattern
event_data = dict(
name=name,
pattern=pattern,
uuid=str(play._uuid),
)
with self.capture_event_data('playbook_on_play_start', **event_data):
super(BaseCallbackModule, self).v2_playbook_on_play_start(play)
def v2_playbook_on_import_for_host(self, result, imported_file):
# NOTE: Not used by Ansible 2.x.
with self.capture_event_data('playbook_on_import_for_host'):
super(BaseCallbackModule, self).v2_playbook_on_import_for_host(result, imported_file)
def v2_playbook_on_not_import_for_host(self, result, missing_file):
# NOTE: Not used by Ansible 2.x.
with self.capture_event_data('playbook_on_not_import_for_host'):
super(BaseCallbackModule, self).v2_playbook_on_not_import_for_host(result, missing_file)
def v2_playbook_on_setup(self):
# NOTE: Not used by Ansible 2.x.
with self.capture_event_data('playbook_on_setup'):
super(BaseCallbackModule, self).v2_playbook_on_setup()
def v2_playbook_on_task_start(self, task, is_conditional):
# FIXME: Flag task path output as vv.
task_uuid = str(task._uuid)
if task_uuid in self.task_uuids:
# FIXME: When this task UUID repeats, it means the play is using the
# free strategy, so different hosts may be running different tasks
# within a play.
return
self.task_uuids.add(task_uuid)
self.set_task(task)
event_data = dict(
task=task,
name=task.get_name(),
is_conditional=is_conditional,
uuid=task_uuid,
)
with self.capture_event_data('playbook_on_task_start', **event_data):
super(BaseCallbackModule, self).v2_playbook_on_task_start(task, is_conditional)
def v2_playbook_on_cleanup_task_start(self, task):
# NOTE: Not used by Ansible 2.x.
self.set_task(task)
event_data = dict(
task=task,
name=task.get_name(),
uuid=str(task._uuid),
is_conditional=True,
)
with self.capture_event_data('playbook_on_task_start', **event_data):
super(BaseCallbackModule, self).v2_playbook_on_cleanup_task_start(task)
def v2_playbook_on_handler_task_start(self, task):
# NOTE: Re-using playbook_on_task_start event for this v2-specific
# event, but setting is_conditional=True, which is how v1 identified a
# task run as a handler.
self.set_task(task)
event_data = dict(
task=task,
name=task.get_name(),
uuid=str(task._uuid),
is_conditional=True,
)
with self.capture_event_data('playbook_on_task_start', **event_data):
super(BaseCallbackModule, self).v2_playbook_on_handler_task_start(task)
def v2_playbook_on_no_hosts_matched(self):
with self.capture_event_data('playbook_on_no_hosts_matched'):
super(BaseCallbackModule, self).v2_playbook_on_no_hosts_matched()
def v2_playbook_on_no_hosts_remaining(self):
with self.capture_event_data('playbook_on_no_hosts_remaining'):
super(BaseCallbackModule, self).v2_playbook_on_no_hosts_remaining()
def v2_playbook_on_notify(self, handler, host):
# NOTE: Not used by Ansible < 2.5.
event_data = dict(
host=host.get_name(),
handler=handler.get_name(),
)
with self.capture_event_data('playbook_on_notify', **event_data):
super(BaseCallbackModule, self).v2_playbook_on_notify(handler, host)
'''
ansible_stats is, retoractively, added in 2.2
'''
def v2_playbook_on_stats(self, stats):
self.clear_play()
# FIXME: Add count of plays/tasks.
event_data = dict(
changed=stats.changed,
dark=stats.dark,
failures=stats.failures,
ok=stats.ok,
processed=stats.processed,
skipped=stats.skipped
)
# write custom set_stat artifact data to the local disk so that it can
# be persisted by awx after the process exits
custom_artifact_data = stats.custom.get('_run', {}) if hasattr(stats, 'custom') else {}
if custom_artifact_data:
# create the directory for custom stats artifacts to live in (if it doesn't exist)
custom_artifacts_dir = os.path.join(os.getenv('AWX_PRIVATE_DATA_DIR'), 'artifacts')
if not os.path.isdir(custom_artifacts_dir):
os.makedirs(custom_artifacts_dir, mode=stat.S_IXUSR + stat.S_IWUSR + stat.S_IRUSR)
custom_artifacts_path = os.path.join(custom_artifacts_dir, 'custom')
with codecs.open(custom_artifacts_path, 'w', encoding='utf-8') as f:
os.chmod(custom_artifacts_path, stat.S_IRUSR | stat.S_IWUSR)
json.dump(custom_artifact_data, f)
with self.capture_event_data('playbook_on_stats', **event_data):
super(BaseCallbackModule, self).v2_playbook_on_stats(stats)
@staticmethod
def _get_event_loop(task):
if hasattr(task, 'loop_with'): # Ansible >=2.5
return task.loop_with
elif hasattr(task, 'loop'): # Ansible <2.4
return task.loop
return None
def v2_runner_on_ok(self, result):
# FIXME: Display detailed results or not based on verbosity.
# strip environment vars from the job event; it already exists on the
# job and sensitive values are filtered there
if result._task.action in ('setup', 'gather_facts'):
result._result.get('ansible_facts', {}).pop('ansible_env', None)
event_data = dict(
host=result._host.get_name(),
remote_addr=result._host.address,
task=result._task,
res=result._result,
event_loop=self._get_event_loop(result._task),
)
with self.capture_event_data('runner_on_ok', **event_data):
super(BaseCallbackModule, self).v2_runner_on_ok(result)
def v2_runner_on_failed(self, result, ignore_errors=False):
# FIXME: Add verbosity for exception/results output.
event_data = dict(
host=result._host.get_name(),
remote_addr=result._host.address,
res=result._result,
task=result._task,
ignore_errors=ignore_errors,
event_loop=self._get_event_loop(result._task),
)
with self.capture_event_data('runner_on_failed', **event_data):
super(BaseCallbackModule, self).v2_runner_on_failed(result, ignore_errors)
def v2_runner_on_skipped(self, result):
event_data = dict(
host=result._host.get_name(),
remote_addr=result._host.address,
task=result._task,
event_loop=self._get_event_loop(result._task),
)
with self.capture_event_data('runner_on_skipped', **event_data):
super(BaseCallbackModule, self).v2_runner_on_skipped(result)
def v2_runner_on_unreachable(self, result):
event_data = dict(
host=result._host.get_name(),
remote_addr=result._host.address,
task=result._task,
res=result._result,
)
with self.capture_event_data('runner_on_unreachable', **event_data):
super(BaseCallbackModule, self).v2_runner_on_unreachable(result)
def v2_runner_on_no_hosts(self, task):
# NOTE: Not used by Ansible 2.x.
event_data = dict(
task=task,
)
with self.capture_event_data('runner_on_no_hosts', **event_data):
super(BaseCallbackModule, self).v2_runner_on_no_hosts(task)
def v2_runner_on_async_poll(self, result):
# NOTE: Not used by Ansible 2.x.
event_data = dict(
host=result._host.get_name(),
task=result._task,
res=result._result,
jid=result._result.get('ansible_job_id'),
)
with self.capture_event_data('runner_on_async_poll', **event_data):
super(BaseCallbackModule, self).v2_runner_on_async_poll(result)
def v2_runner_on_async_ok(self, result):
# NOTE: Not used by Ansible 2.x.
event_data = dict(
host=result._host.get_name(),
task=result._task,
res=result._result,
jid=result._result.get('ansible_job_id'),
)
with self.capture_event_data('runner_on_async_ok', **event_data):
super(BaseCallbackModule, self).v2_runner_on_async_ok(result)
def v2_runner_on_async_failed(self, result):
# NOTE: Not used by Ansible 2.x.
event_data = dict(
host=result._host.get_name(),
task=result._task,
res=result._result,
jid=result._result.get('ansible_job_id'),
)
with self.capture_event_data('runner_on_async_failed', **event_data):
super(BaseCallbackModule, self).v2_runner_on_async_failed(result)
def v2_runner_on_file_diff(self, result, diff):
# NOTE: Not used by Ansible 2.x.
event_data = dict(
host=result._host.get_name(),
task=result._task,
diff=diff,
)
with self.capture_event_data('runner_on_file_diff', **event_data):
super(BaseCallbackModule, self).v2_runner_on_file_diff(result, diff)
def v2_on_file_diff(self, result):
# NOTE: Logged as runner_on_file_diff.
event_data = dict(
host=result._host.get_name(),
task=result._task,
diff=result._result.get('diff'),
)
with self.capture_event_data('runner_on_file_diff', **event_data):
super(BaseCallbackModule, self).v2_on_file_diff(result)
def v2_runner_item_on_ok(self, result):
event_data = dict(
host=result._host.get_name(),
task=result._task,
res=result._result,
)
with self.capture_event_data('runner_item_on_ok', **event_data):
super(BaseCallbackModule, self).v2_runner_item_on_ok(result)
def v2_runner_item_on_failed(self, result):
event_data = dict(
host=result._host.get_name(),
task=result._task,
res=result._result,
)
with self.capture_event_data('runner_item_on_failed', **event_data):
super(BaseCallbackModule, self).v2_runner_item_on_failed(result)
def v2_runner_item_on_skipped(self, result):
event_data = dict(
host=result._host.get_name(),
task=result._task,
res=result._result,
)
with self.capture_event_data('runner_item_on_skipped', **event_data):
super(BaseCallbackModule, self).v2_runner_item_on_skipped(result)
def v2_runner_retry(self, result):
event_data = dict(
host=result._host.get_name(),
task=result._task,
res=result._result,
)
with self.capture_event_data('runner_retry', **event_data):
super(BaseCallbackModule, self).v2_runner_retry(result)
def v2_runner_on_start(self, host, task):
event_data = dict(
host=host.get_name(),
task=task
)
with self.capture_event_data('runner_on_start', **event_data):
super(BaseCallbackModule, self).v2_runner_on_start(host, task)
class AWXDefaultCallbackModule(BaseCallbackModule, DefaultCallbackModule):
CALLBACK_NAME = 'awx_display'
class AWXMinimalCallbackModule(BaseCallbackModule, MinimalCallbackModule):
CALLBACK_NAME = 'minimal'
def v2_playbook_on_play_start(self, play):
pass
def v2_playbook_on_task_start(self, task, is_conditional):
self.set_task(task)

View File

@@ -1,30 +0,0 @@
# Copyright (c) 2017 Ansible by Red Hat
#
# This file is part of Ansible Tower, but depends on code imported from Ansible.
#
# Ansible is free software: you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# Ansible is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with Ansible. If not, see <http://www.gnu.org/licenses/>.
from __future__ import (absolute_import, division, print_function)
# Python
import os
import sys
# Add awx/lib to sys.path.
awx_lib_path = os.path.abspath(os.path.join(os.path.dirname(__file__), '..'))
if awx_lib_path not in sys.path:
sys.path.insert(0, awx_lib_path)
# Tower Display Callback
from awx_display_callback import AWXDefaultCallbackModule as CallbackModule # noqa

View File

@@ -1,30 +0,0 @@
# Copyright (c) 2017 Ansible by Red Hat
#
# This file is part of Ansible Tower, but depends on code imported from Ansible.
#
# Ansible is free software: you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# Ansible is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with Ansible. If not, see <http://www.gnu.org/licenses/>.
from __future__ import (absolute_import, division, print_function)
# Python
import os
import sys
# Add awx/lib to sys.path.
awx_lib_path = os.path.abspath(os.path.join(os.path.dirname(__file__), '..'))
if awx_lib_path not in sys.path:
sys.path.insert(0, awx_lib_path)
# Tower Display Callback
from awx_display_callback import AWXMinimalCallbackModule as CallbackModule # noqa

View File

@@ -1,26 +0,0 @@
# Python
import os
import sys
# Based on http://stackoverflow.com/a/6879344/131141 -- Initialize awx display
# callback as early as possible to wrap ansible.display.Display methods.
def argv_ready(argv):
if argv and os.path.basename(argv[0]) in {'ansible', 'ansible-playbook'}:
import awx_display_callback # noqa
class argv_placeholder(object):
def __del__(self):
try:
argv_ready(sys.argv)
except Exception:
pass
if hasattr(sys, 'argv'):
argv_ready(sys.argv)
else:
sys.argv = argv_placeholder()

View File

@@ -1,2 +0,0 @@
[pytest]
addopts = -v

View File

@@ -1,353 +0,0 @@
# Copyright (c) 2017 Ansible by Red Hat
# All Rights Reserved
from __future__ import absolute_import
from collections import OrderedDict
import json
import os
import shutil
import sys
import tempfile
from unittest import mock
import pytest
# ansible uses `ANSIBLE_CALLBACK_PLUGINS` and `ANSIBLE_STDOUT_CALLBACK` to
# discover callback plugins; `ANSIBLE_CALLBACK_PLUGINS` is a list of paths to
# search for a plugin implementation (which should be named `CallbackModule`)
#
# this code modifies the Python path to make our
# `awx.lib.awx_display_callback` callback importable (because `awx.lib`
# itself is not a package)
#
# we use the `awx_display_callback` imports below within this file, but
# Ansible also uses them when it discovers this file in
# `ANSIBLE_CALLBACK_PLUGINS`
CALLBACK = os.path.splitext(os.path.basename(__file__))[0]
PLUGINS = os.path.dirname(__file__)
with mock.patch.dict(os.environ, {'ANSIBLE_STDOUT_CALLBACK': CALLBACK,
'ANSIBLE_CALLBACK_PLUGINS': PLUGINS}):
from ansible import __version__ as ANSIBLE_VERSION
from ansible.cli.playbook import PlaybookCLI
from ansible.executor.playbook_executor import PlaybookExecutor
from ansible.inventory.manager import InventoryManager
from ansible.parsing.dataloader import DataLoader
from ansible.vars.manager import VariableManager
# Add awx/lib to sys.path so we can use the plugin
path = os.path.abspath(os.path.join(PLUGINS, '..', '..', 'lib'))
if path not in sys.path:
sys.path.insert(0, path)
from awx_display_callback import AWXDefaultCallbackModule as CallbackModule # noqa
from awx_display_callback.events import event_context # noqa
@pytest.fixture()
def cache(request):
class Cache(OrderedDict):
def set(self, key, value):
self[key] = value
local_cache = Cache()
patch = mock.patch.object(event_context, 'cache', local_cache)
patch.start()
request.addfinalizer(patch.stop)
return local_cache
@pytest.fixture()
def executor(tmpdir_factory, request):
playbooks = request.node.callspec.params.get('playbook')
playbook_files = []
for name, playbook in playbooks.items():
filename = str(tmpdir_factory.mktemp('data').join(name))
with open(filename, 'w') as f:
f.write(playbook)
playbook_files.append(filename)
cli = PlaybookCLI(['', 'playbook.yml'])
cli.parse()
options = cli.parser.parse_args(['-v'])[0]
loader = DataLoader()
variable_manager = VariableManager(loader=loader)
inventory = InventoryManager(loader=loader, sources='localhost,')
variable_manager.set_inventory(inventory)
return PlaybookExecutor(playbooks=playbook_files, inventory=inventory,
variable_manager=variable_manager, loader=loader,
options=options, passwords={})
@pytest.mark.parametrize('event', {'playbook_on_start',
'playbook_on_play_start',
'playbook_on_task_start', 'runner_on_ok',
'playbook_on_stats'})
@pytest.mark.parametrize('playbook', [
{'helloworld.yml': '''
- name: Hello World Sample
connection: local
hosts: all
gather_facts: no
tasks:
- name: Hello Message
debug:
msg: "Hello World!"
'''}, # noqa
{'results_included.yml': '''
- name: Run module which generates results list
connection: local
hosts: all
gather_facts: no
vars:
results: ['foo', 'bar']
tasks:
- name: Generate results list
debug:
var: results
'''}, # noqa
])
def test_callback_plugin_receives_events(executor, cache, event, playbook):
executor.run()
assert len(cache)
assert event in [task['event'] for task in cache.values()]
@pytest.mark.parametrize('playbook', [
{'no_log_on_ok.yml': '''
- name: args should not be logged when task-level no_log is set
connection: local
hosts: all
gather_facts: no
tasks:
- shell: echo "SENSITIVE"
no_log: true
'''}, # noqa
{'no_log_on_fail.yml': '''
- name: failed args should not be logged when task-level no_log is set
connection: local
hosts: all
gather_facts: no
tasks:
- shell: echo "SENSITIVE"
no_log: true
failed_when: true
ignore_errors: true
'''}, # noqa
{'no_log_on_skip.yml': '''
- name: skipped task args should be suppressed with no_log
connection: local
hosts: all
gather_facts: no
tasks:
- shell: echo "SENSITIVE"
no_log: true
when: false
'''}, # noqa
{'no_log_on_play.yml': '''
- name: args should not be logged when play-level no_log set
connection: local
hosts: all
gather_facts: no
no_log: true
tasks:
- shell: echo "SENSITIVE"
'''}, # noqa
{'async_no_log.yml': '''
- name: async task args should suppressed with no_log
connection: local
hosts: all
gather_facts: no
no_log: true
tasks:
- async: 10
poll: 1
shell: echo "SENSITIVE"
no_log: true
'''}, # noqa
{'with_items.yml': '''
- name: with_items tasks should be suppressed with no_log
connection: local
hosts: all
gather_facts: no
tasks:
- shell: echo {{ item }}
no_log: true
with_items: [ "SENSITIVE", "SENSITIVE-SKIPPED", "SENSITIVE-FAILED" ]
when: item != "SENSITIVE-SKIPPED"
failed_when: item == "SENSITIVE-FAILED"
ignore_errors: yes
'''}, # noqa, NOTE: with_items will be deprecated in 2.9
{'loop.yml': '''
- name: loop tasks should be suppressed with no_log
connection: local
hosts: all
gather_facts: no
tasks:
- shell: echo {{ item }}
no_log: true
loop: [ "SENSITIVE", "SENSITIVE-SKIPPED", "SENSITIVE-FAILED" ]
when: item != "SENSITIVE-SKIPPED"
failed_when: item == "SENSITIVE-FAILED"
ignore_errors: yes
'''}, # noqa
])
def test_callback_plugin_no_log_filters(executor, cache, playbook):
executor.run()
assert len(cache)
assert 'SENSITIVE' not in json.dumps(cache.items())
@pytest.mark.parametrize('playbook', [
{'no_log_on_ok.yml': '''
- name: args should not be logged when no_log is set at the task or module level
connection: local
hosts: all
gather_facts: no
tasks:
- shell: echo "PUBLIC"
- shell: echo "PRIVATE"
no_log: true
- uri: url=https://example.org username="PUBLIC" password="PRIVATE"
- copy: content="PRIVATE" dest="/tmp/tmp_no_log"
'''}, # noqa
])
def test_callback_plugin_task_args_leak(executor, cache, playbook):
executor.run()
events = cache.values()
assert events[0]['event'] == 'playbook_on_start'
assert events[1]['event'] == 'playbook_on_play_start'
# task 1
assert events[2]['event'] == 'playbook_on_task_start'
assert events[3]['event'] == 'runner_on_ok'
# task 2 no_log=True
assert events[4]['event'] == 'playbook_on_task_start'
assert events[5]['event'] == 'runner_on_ok'
assert 'PUBLIC' in json.dumps(cache.items())
assert 'PRIVATE' not in json.dumps(cache.items())
# make sure playbook was successful, so all tasks were hit
assert not events[-1]['event_data']['failures'], 'Unexpected playbook execution failure'
@pytest.mark.parametrize('playbook', [
{'loop_with_no_log.yml': '''
- name: playbook variable should not be overwritten when using no log
connection: local
hosts: all
gather_facts: no
tasks:
- command: "{{ item }}"
register: command_register
no_log: True
with_items:
- "echo helloworld!"
- debug: msg="{{ command_register.results|map(attribute='stdout')|list }}"
'''}, # noqa
])
def test_callback_plugin_censoring_does_not_overwrite(executor, cache, playbook):
executor.run()
events = cache.values()
assert events[0]['event'] == 'playbook_on_start'
assert events[1]['event'] == 'playbook_on_play_start'
# task 1
assert events[2]['event'] == 'playbook_on_task_start'
# Ordering of task and item events may differ randomly
assert set(['runner_on_ok', 'runner_item_on_ok']) == set([data['event'] for data in events[3:5]])
# task 2 no_log=True
assert events[5]['event'] == 'playbook_on_task_start'
assert events[6]['event'] == 'runner_on_ok'
assert 'helloworld!' in events[6]['event_data']['res']['msg']
@pytest.mark.parametrize('playbook', [
{'strip_env_vars.yml': '''
- name: sensitive environment variables should be stripped from events
connection: local
hosts: all
tasks:
- shell: echo "Hello, World!"
'''}, # noqa
])
def test_callback_plugin_strips_task_environ_variables(executor, cache, playbook):
executor.run()
assert len(cache)
for event in cache.values():
assert os.environ['PATH'] not in json.dumps(event)
@pytest.mark.parametrize('playbook', [
{'custom_set_stat.yml': '''
- name: custom set_stat calls should persist to the local disk so awx can save them
connection: local
hosts: all
tasks:
- set_stats:
data:
foo: "bar"
'''}, # noqa
])
def test_callback_plugin_saves_custom_stats(executor, cache, playbook):
try:
private_data_dir = tempfile.mkdtemp()
with mock.patch.dict(os.environ, {'AWX_PRIVATE_DATA_DIR': private_data_dir}):
executor.run()
artifacts_path = os.path.join(private_data_dir, 'artifacts', 'custom')
with open(artifacts_path, 'r') as f:
assert json.load(f) == {'foo': 'bar'}
finally:
shutil.rmtree(os.path.join(private_data_dir))
@pytest.mark.parametrize('playbook', [
{'handle_playbook_on_notify.yml': '''
- name: handle playbook_on_notify events properly
connection: local
hosts: all
handlers:
- name: my_handler
debug: msg="My Handler"
tasks:
- debug: msg="My Task"
changed_when: true
notify:
- my_handler
'''}, # noqa
])
@pytest.mark.skipif(ANSIBLE_VERSION < '2.5', reason="v2_playbook_on_notify doesn't work before ansible 2.5")
def test_callback_plugin_records_notify_events(executor, cache, playbook):
executor.run()
assert len(cache)
notify_events = [x[1] for x in cache.items() if x[1]['event'] == 'playbook_on_notify']
assert len(notify_events) == 1
assert notify_events[0]['event_data']['handler'] == 'my_handler'
assert notify_events[0]['event_data']['host'] == 'localhost'
assert notify_events[0]['event_data']['task'] == 'debug'
@pytest.mark.parametrize('playbook', [
{'no_log_module_with_var.yml': '''
- name: ensure that module-level secrets are redacted
connection: local
hosts: all
vars:
- pw: SENSITIVE
tasks:
- uri:
url: https://example.org
user: john-jacob-jingleheimer-schmidt
password: "{{ pw }}"
'''}, # noqa
])
def test_module_level_no_log(executor, cache, playbook):
# https://github.com/ansible/tower/issues/1101
# It's possible for `no_log=True` to be defined at the _module_ level,
# e.g., for the URI module password parameter
# This test ensures that we properly redact those
executor.run()
assert len(cache)
assert 'john-jacob-jingleheimer-schmidt' in json.dumps(cache.items())
assert 'SENSITIVE' not in json.dumps(cache.items())

View File

@@ -28,11 +28,19 @@ from awx.main.utils import (
to_python_boolean,
get_licenser,
)
from awx.main.models import * # noqa
from awx.main.models import (
ActivityStream, AdHocCommand, AdHocCommandEvent, Credential, CredentialType,
CredentialInputSource, CustomInventoryScript, Group, Host, Instance, InstanceGroup,
Inventory, InventorySource, InventoryUpdate, InventoryUpdateEvent, Job, JobEvent,
JobHostSummary, JobLaunchConfig, JobTemplate, Label, Notification,
NotificationTemplate, Organization, Project, ProjectUpdate,
ProjectUpdateEvent, Role, Schedule, SystemJob, SystemJobEvent,
SystemJobTemplate, Team, UnifiedJob, UnifiedJobTemplate, WorkflowJob,
WorkflowJobNode, WorkflowJobTemplate, WorkflowJobTemplateNode,
ROLE_SINGLETON_SYSTEM_ADMINISTRATOR, ROLE_SINGLETON_SYSTEM_AUDITOR
)
from awx.main.models.mixins import ResourceMixin
from awx.conf.license import LicenseForbids, feature_enabled
__all__ = ['get_user_queryset', 'check_user_access', 'check_user_access_with_errors',
'user_accessible_objects', 'consumer_access',]
@@ -74,6 +82,17 @@ def get_object_from_data(field, Model, data, obj=None):
raise ParseError(_("Bad data found in related field %s." % field))
def vars_are_encrypted(vars):
'''Returns True if any of the values in the dictionary vars contains
content which is encrypted by the AWX encryption algorithm
'''
for value in vars.values():
if isinstance(value, str):
if value.startswith('$encrypted$'):
return True
return False
def register_access(model_class, access_class):
access_registry[model_class] = access_class
@@ -314,11 +333,35 @@ class BaseAccess(object):
elif not add_host_name and free_instances < 0:
raise PermissionDenied(_("Host count exceeds available instances."))
if feature is not None:
if "features" in validation_info and not validation_info["features"].get(feature, False):
raise LicenseForbids(_("Feature %s is not enabled in the active license.") % feature)
elif "features" not in validation_info:
raise LicenseForbids(_("Features not found in active license."))
def check_org_host_limit(self, data, add_host_name=None):
validation_info = get_licenser().validate()
if validation_info.get('license_type', 'UNLICENSED') == 'open':
return
inventory = get_object_from_data('inventory', Inventory, data)
if inventory is None: # In this case a missing inventory error is launched
return # further down the line, so just ignore it.
org = inventory.organization
if org is None or org.max_hosts == 0:
return
active_count = Host.objects.org_active_count(org.id)
if active_count > org.max_hosts:
raise PermissionDenied(
_("You have already reached the maximum number of %s hosts"
" allowed for your organization. Contact your System Administrator"
" for assistance." % org.max_hosts)
)
if add_host_name:
host_exists = Host.objects.filter(inventory__organization=org.id, name=add_host_name).exists()
if not host_exists and active_count == org.max_hosts:
raise PermissionDenied(
_("You have already reached the maximum number of %s hosts"
" allowed for your organization. Contact your System Administrator"
" for assistance." % org.max_hosts)
)
def get_user_capabilities(self, obj, method_list=[], parent_obj=None, capabilities_cache={}):
if obj is None:
@@ -343,14 +386,11 @@ class BaseAccess(object):
if obj.validation_errors:
user_capabilities[display_method] = False
continue
elif isinstance(obj, (WorkflowJobTemplate, WorkflowJob)) and (not feature_enabled('workflows')):
user_capabilities[display_method] = (display_method == 'delete')
continue
elif display_method == 'copy' and isinstance(obj, WorkflowJobTemplate) and obj.organization_id is None:
user_capabilities[display_method] = self.user.is_superuser
continue
elif display_method == 'copy' and isinstance(obj, Project) and obj.scm_type == '':
# Connot copy manual project without errors
# Cannot copy manual project without errors
user_capabilities[display_method] = False
continue
elif display_method in ['start', 'schedule'] and isinstance(obj, Group): # TODO: remove in 3.3
@@ -386,7 +426,7 @@ class BaseAccess(object):
if display_method == 'schedule':
user_capabilities['schedule'] = user_capabilities['start']
continue
elif display_method == 'delete' and not isinstance(obj, (User, UnifiedJob, CustomInventoryScript)):
elif display_method == 'delete' and not isinstance(obj, (User, UnifiedJob, CustomInventoryScript, CredentialInputSource)):
user_capabilities['delete'] = user_capabilities['edit']
continue
elif display_method == 'copy' and isinstance(obj, (Group, Host)):
@@ -420,6 +460,42 @@ class BaseAccess(object):
return False
class NotificationAttachMixin(BaseAccess):
'''For models that can have notifications attached
I can attach a notification template when
- I have notification_admin_role to organization of the NT
- I can read the object I am attaching it to
I can unattach when those same critiera are met
'''
notification_attach_roles = None
def _can_attach(self, notification_template, resource_obj):
if not NotificationTemplateAccess(self.user).can_change(notification_template, {}):
return False
if self.notification_attach_roles is None:
return self.can_read(resource_obj)
return any(self.user in getattr(resource_obj, role) for role in self.notification_attach_roles)
@check_superuser
def can_attach(self, obj, sub_obj, relationship, data, skip_sub_obj_read_check=False):
if isinstance(sub_obj, NotificationTemplate):
# reverse obj and sub_obj
return self._can_attach(notification_template=sub_obj, resource_obj=obj)
return super(NotificationAttachMixin, self).can_attach(
obj, sub_obj, relationship, data, skip_sub_obj_read_check=skip_sub_obj_read_check)
@check_superuser
def can_unattach(self, obj, sub_obj, relationship, data=None):
if isinstance(sub_obj, NotificationTemplate):
# due to this special case, we use symmetrical logic with attach permission
return self._can_attach(notification_template=sub_obj, resource_obj=obj)
return super(NotificationAttachMixin, self).can_unattach(
obj, sub_obj, relationship, relationship, data=data
)
class InstanceAccess(BaseAccess):
model = Instance
@@ -434,12 +510,16 @@ class InstanceAccess(BaseAccess):
skip_sub_obj_read_check=False):
if relationship == 'rampart_groups' and isinstance(sub_obj, InstanceGroup):
return self.user.is_superuser
return super(InstanceAccess, self).can_attach(obj, sub_obj, relationship, *args, **kwargs)
return super(InstanceAccess, self).can_attach(
obj, sub_obj, relationship, data, skip_sub_obj_read_check=skip_sub_obj_read_check
)
def can_unattach(self, obj, sub_obj, relationship, data=None):
if relationship == 'rampart_groups' and isinstance(sub_obj, InstanceGroup):
return self.user.is_superuser
return super(InstanceAccess, self).can_unattach(obj, sub_obj, relationship, *args, **kwargs)
return super(InstanceAccess, self).can_unattach(
obj, sub_obj, relationship, relationship, data=data
)
def can_add(self, data):
return False
@@ -576,23 +656,22 @@ class UserAccess(BaseAccess):
return False
def can_attach(self, obj, sub_obj, relationship, *args, **kwargs):
if not settings.MANAGE_ORGANIZATION_AUTH and not self.user.is_superuser:
return False
# Reverse obj and sub_obj, defer to RoleAccess if this is a role assignment.
# The only thing that a User should ever have attached is a Role
if relationship == 'roles':
role_access = RoleAccess(self.user)
return role_access.can_attach(sub_obj, obj, 'members', *args, **kwargs)
return super(UserAccess, self).can_attach(obj, sub_obj, relationship, *args, **kwargs)
logger.error('Unexpected attempt to associate {} with a user.'.format(sub_obj))
return False
def can_unattach(self, obj, sub_obj, relationship, *args, **kwargs):
if not settings.MANAGE_ORGANIZATION_AUTH and not self.user.is_superuser:
return False
# The only thing that a User should ever have to be unattached is a Role
if relationship == 'roles':
role_access = RoleAccess(self.user)
return role_access.can_unattach(sub_obj, obj, 'members', *args, **kwargs)
return super(UserAccess, self).can_unattach(obj, sub_obj, relationship, *args, **kwargs)
logger.error('Unexpected attempt to de-associate {} from a user.'.format(sub_obj))
return False
class OAuth2ApplicationAccess(BaseAccess):
@@ -608,13 +687,14 @@ class OAuth2ApplicationAccess(BaseAccess):
model = OAuth2Application
select_related = ('user',)
prefetch_related = ('organization', 'oauth2accesstoken_set')
def filtered_queryset(self):
org_access_qs = Organization.accessible_objects(self.user, 'member_role')
return self.model.objects.filter(organization__in=org_access_qs)
def can_change(self, obj, data):
return self.user.is_superuser or self.check_related('organization', Organization, data, obj=obj,
return self.user.is_superuser or self.check_related('organization', Organization, data, obj=obj,
role_field='admin_role', mandatory=True)
def can_delete(self, obj):
@@ -622,7 +702,7 @@ class OAuth2ApplicationAccess(BaseAccess):
def can_add(self, data):
if self.user.is_superuser:
return True
return True
if not data:
return Organization.accessible_objects(self.user, 'admin_role').exists()
return self.check_related('organization', Organization, data, role_field='admin_role', mandatory=True)
@@ -636,29 +716,30 @@ class OAuth2TokenAccess(BaseAccess):
- I am the user of the token.
I can create an OAuth2 app token when:
- I have the read permission of the related application.
I can read, change or delete a personal token when:
I can read, change or delete a personal token when:
- I am the user of the token
- I am the superuser
I can create an OAuth2 Personal Access Token when:
- I am a user. But I can only create a PAT for myself.
- I am a user. But I can only create a PAT for myself.
'''
model = OAuth2AccessToken
select_related = ('user', 'application')
def filtered_queryset(self):
prefetch_related = ('refresh_token',)
def filtered_queryset(self):
org_access_qs = Organization.objects.filter(
Q(admin_role__members=self.user) | Q(auditor_role__members=self.user))
return self.model.objects.filter(application__organization__in=org_access_qs) | self.model.objects.filter(user__id=self.user.pk)
def can_delete(self, obj):
if (self.user.is_superuser) | (obj.user == self.user):
return True
elif not obj.application:
return False
return self.user in obj.application.organization.admin_role
def can_change(self, obj, data):
return self.can_delete(obj)
@@ -671,7 +752,7 @@ class OAuth2TokenAccess(BaseAccess):
return True
class OrganizationAccess(BaseAccess):
class OrganizationAccess(NotificationAttachMixin, BaseAccess):
'''
I can see organizations when:
- I am a superuser.
@@ -685,6 +766,8 @@ class OrganizationAccess(BaseAccess):
model = Organization
prefetch_related = ('created_by', 'modified_by',)
# organization admin_role is not a parent of organization auditor_role
notification_attach_roles = ['admin_role', 'auditor_role']
def filtered_queryset(self):
return self.model.accessible_objects(self.user, 'read_role')
@@ -694,13 +777,18 @@ class OrganizationAccess(BaseAccess):
return self.user in obj.admin_role
def can_delete(self, obj):
self.check_license(feature='multiple_organizations', check_expiration=False)
self.check_license(check_expiration=False)
is_change_possible = self.can_change(obj, None)
if not is_change_possible:
return False
return True
def can_attach(self, obj, sub_obj, relationship, *args, **kwargs):
# If the request is updating the membership, check the membership role permissions instead
if relationship in ('member_role.members', 'admin_role.members'):
rel_role = getattr(obj, relationship.split('.')[0])
return RoleAccess(self.user).can_attach(rel_role, sub_obj, 'members', *args, **kwargs)
if relationship == "instance_groups":
if self.user.is_superuser:
return True
@@ -708,6 +796,11 @@ class OrganizationAccess(BaseAccess):
return super(OrganizationAccess, self).can_attach(obj, sub_obj, relationship, *args, **kwargs)
def can_unattach(self, obj, sub_obj, relationship, *args, **kwargs):
# If the request is updating the membership, check the membership role permissions instead
if relationship in ('member_role.members', 'admin_role.members'):
rel_role = getattr(obj, relationship.split('.')[0])
return RoleAccess(self.user).can_unattach(rel_role, sub_obj, 'members', *args, **kwargs)
if relationship == "instance_groups":
return self.can_attach(obj, sub_obj, relationship, *args, **kwargs)
return super(OrganizationAccess, self).can_attach(obj, sub_obj, relationship, *args, **kwargs)
@@ -735,7 +828,7 @@ class InventoryAccess(BaseAccess):
'''
model = Inventory
select_related = ('created_by', 'modified_by', 'organization',)
prefetch_related = ('created_by', 'modified_by', 'organization')
def filtered_queryset(self, allowed=None, ad_hoc=None):
return self.model.accessible_objects(self.user, 'read_role')
@@ -826,6 +919,10 @@ class HostAccess(BaseAccess):
# Check to see if we have enough licenses
self.check_license(add_host_name=data.get('name', None))
# Check the per-org limit
self.check_org_host_limit(data, add_host_name=data.get('name', None))
return True
def can_change(self, obj, data):
@@ -838,6 +935,10 @@ class HostAccess(BaseAccess):
if data and 'name' in data:
self.check_license(add_host_name=data['name'])
# Check the per-org limit
self.check_org_host_limit({'inventory': obj.inventory},
add_host_name=data['name'])
# Checks for admin or change permission on inventory, controls whether
# the user can edit variable data.
return obj and self.user in obj.inventory.admin_role
@@ -900,21 +1001,8 @@ class GroupAccess(BaseAccess):
def can_delete(self, obj):
return bool(obj and self.user in obj.inventory.admin_role)
def can_start(self, obj, validate_license=True):
# TODO: Delete for 3.3, only used by v1 serializer
# Used as another alias to inventory_source start access for user_capabilities
if obj:
try:
return self.user.can_access(
InventorySource, 'start', obj.deprecated_inventory_source,
validate_license=validate_license)
obj.deprecated_inventory_source
except Group.deprecated_inventory_source.RelatedObjectDoesNotExist:
return False
return False
class InventorySourceAccess(BaseAccess):
class InventorySourceAccess(NotificationAttachMixin, BaseAccess):
'''
I can see inventory sources whenever I can see their inventory.
I can change inventory sources whenever I can change their inventory.
@@ -991,8 +1079,8 @@ class InventoryUpdateAccess(BaseAccess):
'''
model = InventoryUpdate
select_related = ('created_by', 'modified_by', 'inventory_source__inventory',)
prefetch_related = ('unified_job_template', 'instance_group', 'credentials',)
select_related = ('created_by', 'modified_by', 'inventory_source',)
prefetch_related = ('unified_job_template', 'instance_group', 'credentials__credential_type', 'inventory', 'source_script')
def filtered_queryset(self):
return self.model.objects.filter(inventory_source__inventory__in=Inventory.accessible_pk_qs(self.user, 'read_role'))
@@ -1006,11 +1094,7 @@ class InventoryUpdateAccess(BaseAccess):
return self.user in obj.inventory_source.inventory.admin_role
def can_start(self, obj, validate_license=True):
# For relaunching
if obj and obj.inventory_source:
access = InventorySourceAccess(self.user)
return access.can_start(obj.inventory_source, validate_license=validate_license)
return False
return InventorySourceAccess(self.user).can_start(obj, validate_license=validate_license)
@check_superuser
def can_delete(self, obj):
@@ -1028,6 +1112,7 @@ class CredentialTypeAccess(BaseAccess):
'''
model = CredentialType
prefetch_related = ('created_by', 'modified_by',)
def can_read(self, obj):
return True
@@ -1110,6 +1195,55 @@ class CredentialAccess(BaseAccess):
# return True
return self.can_change(obj, None)
def get_user_capabilities(self, obj, **kwargs):
user_capabilities = super(CredentialAccess, self).get_user_capabilities(obj, **kwargs)
user_capabilities['use'] = self.can_use(obj)
return user_capabilities
class CredentialInputSourceAccess(BaseAccess):
'''
I can see a CredentialInputSource when:
- I can see the associated target_credential
I can create/change a CredentialInputSource when:
- I'm an admin of the associated target_credential
- I have use access to the associated source credential
I can delete a CredentialInputSource when:
- I'm an admin of the associated target_credential
'''
model = CredentialInputSource
select_related = ('target_credential', 'source_credential')
def filtered_queryset(self):
return CredentialInputSource.objects.filter(
target_credential__in=Credential.accessible_pk_qs(self.user, 'read_role'))
@check_superuser
def can_read(self, obj):
return self.user in obj.target_credential.read_role
@check_superuser
def can_add(self, data):
return (
self.check_related('target_credential', Credential, data, role_field='admin_role') and
self.check_related('source_credential', Credential, data, role_field='use_role')
)
@check_superuser
def can_change(self, obj, data):
if self.can_add(data) is False:
return False
return (
self.user in obj.target_credential.admin_role and
self.user in obj.source_credential.use_role
)
@check_superuser
def can_delete(self, obj):
return self.user in obj.target_credential.admin_role
class TeamAccess(BaseAccess):
'''
@@ -1117,6 +1251,7 @@ class TeamAccess(BaseAccess):
- I'm a superuser.
- I'm an admin of the team
- I'm a member of that team.
- I'm a member of the team's organization
I can create/change a team when:
- I'm a superuser.
- I'm an admin for the team
@@ -1129,7 +1264,10 @@ class TeamAccess(BaseAccess):
if settings.ORG_ADMINS_CAN_SEE_ALL_USERS and \
(self.user.admin_of_organizations.exists() or self.user.auditor_of_organizations.exists()):
return self.model.objects.all()
return self.model.accessible_objects(self.user, 'read_role')
return self.model.objects.filter(
Q(organization=Organization.accessible_pk_qs(self.user, 'member_role')) |
Q(pk__in=self.model.accessible_pk_qs(self.user, 'read_role'))
)
@check_superuser
def can_add(self, data):
@@ -1167,6 +1305,12 @@ class TeamAccess(BaseAccess):
*args, **kwargs)
if self.user.is_superuser:
return True
# If the request is updating the membership, check the membership role permissions instead
if relationship in ('member_role.members', 'admin_role.members'):
rel_role = getattr(obj, relationship.split('.')[0])
return RoleAccess(self.user).can_attach(rel_role, sub_obj, 'members', *args, **kwargs)
return super(TeamAccess, self).can_attach(obj, sub_obj, relationship,
*args, **kwargs)
@@ -1177,11 +1321,17 @@ class TeamAccess(BaseAccess):
role_access = RoleAccess(self.user)
return role_access.can_unattach(sub_obj, obj, 'member_role.parents',
*args, **kwargs)
# If the request is updating the membership, check the membership role permissions instead
if relationship in ('member_role.members', 'admin_role.members'):
rel_role = getattr(obj, relationship.split('.')[0])
return RoleAccess(self.user).can_unattach(rel_role, sub_obj, 'members', *args, **kwargs)
return super(TeamAccess, self).can_unattach(obj, sub_obj, relationship,
*args, **kwargs)
class ProjectAccess(BaseAccess):
class ProjectAccess(NotificationAttachMixin, BaseAccess):
'''
I can see projects when:
- I am a superuser.
@@ -1199,7 +1349,9 @@ class ProjectAccess(BaseAccess):
'''
model = Project
select_related = ('modified_by', 'credential', 'current_job', 'last_job',)
select_related = ('credential',)
prefetch_related = ('modified_by', 'created_by', 'organization', 'last_job', 'current_job')
notification_attach_roles = ['admin_role']
def filtered_queryset(self):
return self.model.accessible_objects(self.user, 'read_role')
@@ -1262,7 +1414,7 @@ class ProjectUpdateAccess(BaseAccess):
return obj and self.user in obj.project.admin_role
class JobTemplateAccess(BaseAccess):
class JobTemplateAccess(NotificationAttachMixin, BaseAccess):
'''
I can see job templates when:
- I have read role for the job template.
@@ -1332,7 +1484,7 @@ class JobTemplateAccess(BaseAccess):
return self.user in project.use_role
else:
return False
@check_superuser
def can_copy_related(self, obj):
'''
@@ -1341,17 +1493,19 @@ class JobTemplateAccess(BaseAccess):
'''
# obj.credentials.all() is accessible ONLY when object is saved (has valid id)
credential_manager = getattr(obj, 'credentials', None) if getattr(obj, 'id', False) else Credentials.objects.none()
return reduce(lambda prev, cred: prev and self.user in cred.use_role, credential_manager.all(), True)
credential_manager = getattr(obj, 'credentials', None) if getattr(obj, 'id', False) else Credential.objects.none()
user_can_copy = reduce(lambda prev, cred: prev and self.user in cred.use_role, credential_manager.all(), True)
if not user_can_copy:
raise PermissionDenied(_('Insufficient access to Job Template credentials.'))
return user_can_copy
def can_start(self, obj, validate_license=True):
# Check license.
if validate_license:
self.check_license()
if obj.survey_enabled:
self.check_license(feature='surveys')
if Instance.objects.active_count() > 1:
self.check_license(feature='ha')
# Check the per-org limit
self.check_org_host_limit({'inventory': obj.inventory})
# Super users can start any job
if self.user.is_superuser:
@@ -1409,8 +1563,6 @@ class JobTemplateAccess(BaseAccess):
@check_superuser
def can_attach(self, obj, sub_obj, relationship, data, skip_sub_obj_read_check=False):
if isinstance(sub_obj, NotificationTemplate):
return self.check_related('organization', Organization, {}, obj=sub_obj, mandatory=True)
if relationship == "instance_groups":
if not obj.project.organization:
return False
@@ -1506,6 +1658,9 @@ class JobAccess(BaseAccess):
if validate_license:
self.check_license()
# Check the per-org limit
self.check_org_host_limit({'inventory': obj.inventory})
# A super user can relaunch a job
if self.user.is_superuser:
return True
@@ -1523,10 +1678,10 @@ class JobAccess(BaseAccess):
prompts_access = False
elif not config.has_user_prompts(obj.job_template):
prompts_access = True
elif obj.created_by_id != self.user.pk:
elif obj.created_by_id != self.user.pk and vars_are_encrypted(config.extra_data):
prompts_access = False
if self.save_messages:
self.messages['detail'] = _('Job was launched with prompts provided by another user.')
self.messages['detail'] = _('Job was launched with secret prompts provided by another user.')
else:
prompts_access = (
JobLaunchConfigAccess(self.user).can_add({'reference_obj': config}) and
@@ -1688,7 +1843,7 @@ class WorkflowJobTemplateNodeAccess(BaseAccess):
'''
model = WorkflowJobTemplateNode
prefetch_related = ('success_nodes', 'failure_nodes', 'always_nodes',
'unified_job_template', 'credentials',)
'unified_job_template', 'credentials', 'workflow_job_template')
def filtered_queryset(self):
return self.model.objects.filter(
@@ -1780,9 +1935,8 @@ class WorkflowJobNodeAccess(BaseAccess):
Deletion must happen as a cascade delete from the workflow job.
'''
model = WorkflowJobNode
select_related = ('unified_job_template', 'job',)
prefetch_related = ('success_nodes', 'failure_nodes', 'always_nodes',
'credentials',)
prefetch_related = ('unified_job_template', 'job', 'workflow_job', 'credentials',
'success_nodes', 'failure_nodes', 'always_nodes',)
def filtered_queryset(self):
return self.model.objects.filter(
@@ -1805,7 +1959,7 @@ class WorkflowJobNodeAccess(BaseAccess):
# TODO: notification attachments?
class WorkflowJobTemplateAccess(BaseAccess):
class WorkflowJobTemplateAccess(NotificationAttachMixin, BaseAccess):
'''
I can only see/manage Workflow Job Templates if I'm a super user
'''
@@ -1850,7 +2004,6 @@ class WorkflowJobTemplateAccess(BaseAccess):
qs = obj.workflow_job_template_nodes
qs = qs.prefetch_related('unified_job_template', 'inventory__use_role', 'credentials__use_role')
for node in qs.all():
node_errors = {}
if node.inventory and self.user not in node.inventory.use_role:
missing_inventories.append(node.inventory.name)
for cred in node.credentials.all():
@@ -1859,8 +2012,6 @@ class WorkflowJobTemplateAccess(BaseAccess):
ujt = node.unified_job_template
if ujt and not self.user.can_access(UnifiedJobTemplate, 'start', ujt, validate_license=False):
missing_ujt.append(ujt.name)
if node_errors:
wfjt_errors[node.id] = node_errors
if missing_ujt:
self.messages['templates_unable_to_copy'] = missing_ujt
if missing_credentials:
@@ -1875,9 +2026,9 @@ class WorkflowJobTemplateAccess(BaseAccess):
if validate_license:
# check basic license, node count
self.check_license()
# if surveys are added to WFJTs, check license here
if obj.survey_enabled:
self.check_license(feature='surveys')
# Check the per-org limit
self.check_org_host_limit({'inventory': obj.inventory})
# Super users can start any job
if self.user.is_superuser:
@@ -1886,11 +2037,6 @@ class WorkflowJobTemplateAccess(BaseAccess):
return self.user in obj.execute_role
def can_change(self, obj, data):
# Check survey license if surveys are added to WFJTs
if (data and 'survey_enabled' in data and
obj.survey_enabled != data['survey_enabled'] and data['survey_enabled']):
self.check_license(feature='surveys')
if self.user.is_superuser:
return True
@@ -1946,6 +2092,9 @@ class WorkflowJobAccess(BaseAccess):
if validate_license:
self.check_license()
# Check the per-org limit
self.check_org_host_limit({'inventory': obj.inventory})
if self.user.is_superuser:
return True
@@ -1967,9 +2116,9 @@ class WorkflowJobAccess(BaseAccess):
# Check if access to prompts to prevent relaunch
if config.prompts_dict():
if obj.created_by_id != self.user.pk:
if obj.created_by_id != self.user.pk and vars_are_encrypted(config.extra_data):
if self.save_messages:
self.messages['detail'] = _('Job was launched with prompts provided by another user.')
self.messages['detail'] = _('Job was launched with secret prompts provided by another user.')
return False
if not JobLaunchConfigAccess(self.user).can_add({'reference_obj': config}):
if self.save_messages:
@@ -2022,6 +2171,9 @@ class AdHocCommandAccess(BaseAccess):
if validate_license:
self.check_license()
# Check the per-org limit
self.check_org_host_limit(data)
# If a credential is provided, the user should have use access to it.
if not self.check_related('credential', Credential, data, role_field='use_role'):
return False
@@ -2112,7 +2264,7 @@ class JobEventAccess(BaseAccess):
'''
model = JobEvent
prefetch_related = ('hosts', 'children', 'job__job_template', 'host',)
prefetch_related = ('hosts', 'job__job_template', 'host',)
def filtered_queryset(self):
return self.model.objects.filter(
@@ -2222,11 +2374,6 @@ class UnifiedJobTemplateAccess(BaseAccess):
Q(inventorysource__inventory__id__in=Inventory._accessible_pk_qs(
Inventory, self.user, 'read_role')))
def get_queryset(self):
# TODO: remove after the depreciation of v1 API
qs = super(UnifiedJobTemplateAccess, self).get_queryset()
return qs.exclude(inventorysource__source="")
def can_start(self, obj, validate_license=True):
access_class = access_registry[obj.__class__]
access_instance = access_class(self.user)
@@ -2331,6 +2478,7 @@ class NotificationTemplateAccess(BaseAccess):
I can see/use a notification_template if I have permission to
'''
model = NotificationTemplate
prefetch_related = ('created_by', 'modified_by', 'organization')
def filtered_queryset(self):
return self.model.objects.filter(
@@ -2431,7 +2579,7 @@ class ActivityStreamAccess(BaseAccess):
model = ActivityStream
prefetch_related = ('organization', 'user', 'inventory', 'host', 'group',
'inventory_update', 'credential', 'credential_type', 'team',
'ad_hoc_command', 'o_auth2_application', 'o_auth2_access_token',
'ad_hoc_command', 'o_auth2_application', 'o_auth2_access_token',
'notification_template', 'notification', 'label', 'role', 'actor',
'schedule', 'custom_inventory_script', 'unified_job_template',
'workflow_job_template_node',)
@@ -2517,6 +2665,7 @@ class ActivityStreamAccess(BaseAccess):
class CustomInventoryScriptAccess(BaseAccess):
model = CustomInventoryScript
prefetch_related = ('created_by', 'modified_by', 'organization')
def filtered_queryset(self):
return self.model.accessible_objects(self.user, 'read_role').all()
@@ -2550,6 +2699,17 @@ class RoleAccess(BaseAccess):
'''
model = Role
prefetch_related = ('content_type',)
def filtered_queryset(self):
result = Role.visible_roles(self.user)
# Sanity check: is the requesting user an orphaned non-admin/auditor?
# if yes, make system admin/auditor mandatorily visible.
if not self.user.is_superuser and not self.user.is_system_auditor and not self.user.organizations.exists():
mandatories = ('system_administrator', 'system_auditor')
super_qs = Role.objects.filter(singleton_name__in=mandatories)
result = result | super_qs
return result
def can_read(self, obj):
if not obj:
@@ -2569,10 +2729,6 @@ class RoleAccess(BaseAccess):
@check_superuser
def can_unattach(self, obj, sub_obj, relationship, data=None, skip_sub_obj_read_check=False):
if isinstance(obj.content_object, Team):
if not settings.MANAGE_ORGANIZATION_AUTH and not self.user.is_superuser:
return False
if not skip_sub_obj_read_check and relationship in ['members', 'member_role.parents', 'parents']:
# If we are unattaching a team Role, check the Team read access
if relationship == 'parents':
@@ -2584,18 +2740,22 @@ class RoleAccess(BaseAccess):
# Being a user in the member_role or admin_role of an organization grants
# administrators of that Organization the ability to edit that user. To prevent
# unwanted escalations lets ensure that the Organization administartor has the abilty
# unwanted escalations let's ensure that the Organization administrator has the ability
# to admin the user being added to the role.
if (isinstance(obj.content_object, Organization) and
obj.role_field in (Organization.member_role.field.parent_role + ['member_role'])):
if isinstance(obj.content_object, Organization) and obj.role_field in ['admin_role', 'member_role']:
if not isinstance(sub_obj, User):
logger.error('Unexpected attempt to associate {} with organization role.'.format(sub_obj))
return False
if not settings.MANAGE_ORGANIZATION_AUTH and not self.user.is_superuser:
return False
if not UserAccess(self.user).can_admin(sub_obj, None, allow_orphans=True):
return False
if isinstance(obj.content_object, ResourceMixin) and \
self.user in obj.content_object.admin_role:
if isinstance(obj.content_object, Team) and obj.role_field in ['admin_role', 'member_role']:
if not settings.MANAGE_ORGANIZATION_AUTH and not self.user.is_superuser:
return False
if isinstance(obj.content_object, ResourceMixin) and self.user in obj.content_object.admin_role:
return True
return False

View File

@@ -0,0 +1 @@
from .core import register, gather, ship # noqa

View File

@@ -0,0 +1,276 @@
import os
import os.path
import platform
from django.db import connection
from django.db.models import Count
from django.conf import settings
from django.utils.timezone import now
from awx.conf.license import get_license
from awx.main.utils import (get_awx_version, get_ansible_version,
get_custom_venv_choices, camelcase_to_underscore)
from awx.main import models
from django.contrib.sessions.models import Session
from awx.main.analytics import register
'''
This module is used to define metrics collected by awx.main.analytics.gather()
Each function is decorated with a key name, and should return a data
structure that can be serialized to JSON
@register('something')
def something(since):
# the generated archive will contain a `something.json` w/ this JSON
return {'some': 'json'}
All functions - when called - will be passed a datetime.datetime object,
`since`, which represents the last time analytics were gathered (some metrics
functions - like those that return metadata about playbook runs, may return
data _since_ the last report date - i.e., new data in the last 24 hours)
'''
@register('config')
def config(since):
license_info = get_license(show_key=False)
install_type = 'traditional'
if os.environ.get('container') == 'oci':
install_type = 'openshift'
elif 'KUBERNETES_SERVICE_PORT' in os.environ:
install_type = 'k8s'
return {
'platform': {
'system': platform.system(),
'dist': platform.dist(),
'release': platform.release(),
'type': install_type,
},
'install_uuid': settings.INSTALL_UUID,
'instance_uuid': settings.SYSTEM_UUID,
'tower_url_base': settings.TOWER_URL_BASE,
'tower_version': get_awx_version(),
'ansible_version': get_ansible_version(),
'license_type': license_info.get('license_type', 'UNLICENSED'),
'free_instances': license_info.get('free instances', 0),
'license_expiry': license_info.get('time_remaining', 0),
'pendo_tracking': settings.PENDO_TRACKING_STATE,
'authentication_backends': settings.AUTHENTICATION_BACKENDS,
'logging_aggregators': settings.LOG_AGGREGATOR_LOGGERS,
'external_logger_enabled': settings.LOG_AGGREGATOR_ENABLED,
'external_logger_type': getattr(settings, 'LOG_AGGREGATOR_TYPE', None),
}
@register('counts')
def counts(since):
counts = {}
for cls in (models.Organization, models.Team, models.User,
models.Inventory, models.Credential, models.Project,
models.JobTemplate, models.WorkflowJobTemplate,
models.Host, models.Schedule, models.CustomInventoryScript,
models.NotificationTemplate):
counts[camelcase_to_underscore(cls.__name__)] = cls.objects.count()
venvs = get_custom_venv_choices()
counts['custom_virtualenvs'] = len([
v for v in venvs
if os.path.basename(v.rstrip('/')) != 'ansible'
])
inv_counts = dict(models.Inventory.objects.order_by().values_list('kind').annotate(Count('kind')))
inv_counts['normal'] = inv_counts.get('', 0)
inv_counts.pop('', None)
inv_counts['smart'] = inv_counts.get('smart', 0)
counts['inventories'] = inv_counts
counts['unified_job'] = models.UnifiedJob.objects.exclude(launch_type='sync').count() # excludes implicit project_updates
counts['active_host_count'] = models.Host.objects.active_count()
active_sessions = Session.objects.filter(expire_date__gte=now()).count()
active_user_sessions = models.UserSessionMembership.objects.select_related('session').filter(session__expire_date__gte=now()).count()
active_anonymous_sessions = active_sessions - active_user_sessions
counts['active_sessions'] = active_sessions
counts['active_user_sessions'] = active_user_sessions
counts['active_anonymous_sessions'] = active_anonymous_sessions
counts['running_jobs'] = models.UnifiedJob.objects.exclude(launch_type='sync').filter(status__in=('running', 'waiting',)).count()
return counts
@register('org_counts')
def org_counts(since):
counts = {}
for org in models.Organization.objects.annotate(num_users=Count('member_role__members', distinct=True),
num_teams=Count('teams', distinct=True)).values('name', 'id', 'num_users', 'num_teams'):
counts[org['id']] = {'name': org['name'],
'users': org['num_users'],
'teams': org['num_teams']
}
return counts
@register('cred_type_counts')
def cred_type_counts(since):
counts = {}
for cred_type in models.CredentialType.objects.annotate(num_credentials=Count(
'credentials', distinct=True)).values('name', 'id', 'managed_by_tower', 'num_credentials'):
counts[cred_type['id']] = {'name': cred_type['name'],
'credential_count': cred_type['num_credentials'],
'managed_by_tower': cred_type['managed_by_tower']
}
return counts
@register('inventory_counts')
def inventory_counts(since):
counts = {}
for inv in models.Inventory.objects.filter(kind='').annotate(num_sources=Count('inventory_sources', distinct=True),
num_hosts=Count('hosts', distinct=True)).only('id', 'name', 'kind'):
counts[inv.id] = {'name': inv.name,
'kind': inv.kind,
'hosts': inv.num_hosts,
'sources': inv.num_sources
}
for smart_inv in models.Inventory.objects.filter(kind='smart'):
counts[smart_inv.id] = {'name': smart_inv.name,
'kind': smart_inv.kind,
'num_hosts': smart_inv.hosts.count(),
'num_sources': smart_inv.inventory_sources.count()
}
return counts
@register('projects_by_scm_type')
def projects_by_scm_type(since):
counts = dict(
(t[0] or 'manual', 0)
for t in models.Project.SCM_TYPE_CHOICES
)
for result in models.Project.objects.values('scm_type').annotate(
count=Count('scm_type')
).order_by('scm_type'):
counts[result['scm_type'] or 'manual'] = result['count']
return counts
@register('instance_info')
def instance_info(since):
info = {}
instances = models.Instance.objects.values_list('hostname').annotate().values(
'uuid', 'version', 'capacity', 'cpu', 'memory', 'managed_by_policy', 'hostname', 'last_isolated_check', 'enabled')
for instance in instances:
instance_info = {
'uuid': instance['uuid'],
'version': instance['version'],
'capacity': instance['capacity'],
'cpu': instance['cpu'],
'memory': instance['memory'],
'managed_by_policy': instance['managed_by_policy'],
'last_isolated_check': instance['last_isolated_check'],
'enabled': instance['enabled']
}
info[instance['uuid']] = instance_info
return info
@register('job_counts')
def job_counts(since):
counts = {}
counts['total_jobs'] = models.UnifiedJob.objects.exclude(launch_type='sync').count()
counts['status'] = dict(models.UnifiedJob.objects.exclude(launch_type='sync').values_list('status').annotate(Count('status')).order_by())
counts['launch_type'] = dict(models.UnifiedJob.objects.exclude(launch_type='sync').values_list(
'launch_type').annotate(Count('launch_type')).order_by())
return counts
@register('job_instance_counts')
def job_instance_counts(since):
counts = {}
job_types = models.UnifiedJob.objects.exclude(launch_type='sync').values_list(
'execution_node', 'launch_type').annotate(job_launch_type=Count('launch_type'))
for job in job_types:
counts.setdefault(job[0], {}).setdefault('launch_type', {})[job[1]] = job[2]
job_statuses = models.UnifiedJob.objects.exclude(launch_type='sync').values_list(
'execution_node', 'status').annotate(job_status=Count('status'))
for job in job_statuses:
counts.setdefault(job[0], {}).setdefault('status', {})[job[1]] = job[2]
return counts
# Copies Job Events from db to a .csv to be shipped
def copy_tables(since, full_path):
def _copy_table(table, query, path):
file_path = os.path.join(path, table + '_table.csv')
file = open(file_path, 'w', encoding='utf-8')
with connection.cursor() as cursor:
cursor.copy_expert(query, file)
file.close()
return file_path
events_query = '''COPY (SELECT main_jobevent.id,
main_jobevent.created,
main_jobevent.uuid,
main_jobevent.parent_uuid,
main_jobevent.event,
main_jobevent.event_data::json->'task_action' AS task_action,
main_jobevent.failed,
main_jobevent.changed,
main_jobevent.playbook,
main_jobevent.play,
main_jobevent.task,
main_jobevent.role,
main_jobevent.job_id,
main_jobevent.host_id,
main_jobevent.host_name
FROM main_jobevent
WHERE main_jobevent.created > {}
ORDER BY main_jobevent.id ASC) TO STDOUT WITH CSV HEADER'''.format(since.strftime("'%Y-%m-%d %H:%M:%S'"))
_copy_table(table='events', query=events_query, path=full_path)
unified_job_query = '''COPY (SELECT main_unifiedjob.id,
main_unifiedjob.polymorphic_ctype_id,
django_content_type.model,
main_unifiedjob.created,
main_unifiedjob.name,
main_unifiedjob.unified_job_template_id,
main_unifiedjob.launch_type,
main_unifiedjob.schedule_id,
main_unifiedjob.execution_node,
main_unifiedjob.controller_node,
main_unifiedjob.cancel_flag,
main_unifiedjob.status,
main_unifiedjob.failed,
main_unifiedjob.started,
main_unifiedjob.finished,
main_unifiedjob.elapsed,
main_unifiedjob.job_explanation,
main_unifiedjob.instance_group_id
FROM main_unifiedjob, django_content_type
WHERE main_unifiedjob.created > {} AND
main_unifiedjob.polymorphic_ctype_id = django_content_type.id AND
main_unifiedjob.launch_type != 'sync'
ORDER BY main_unifiedjob.id ASC) TO STDOUT WITH CSV HEADER'''.format(since.strftime("'%Y-%m-%d %H:%M:%S'"))
_copy_table(table='unified_jobs', query=unified_job_query, path=full_path)
unified_job_template_query = '''COPY (SELECT main_unifiedjobtemplate.id,
main_unifiedjobtemplate.polymorphic_ctype_id,
django_content_type.model,
main_unifiedjobtemplate.created,
main_unifiedjobtemplate.modified,
main_unifiedjobtemplate.created_by_id,
main_unifiedjobtemplate.modified_by_id,
main_unifiedjobtemplate.name,
main_unifiedjobtemplate.current_job_id,
main_unifiedjobtemplate.last_job_id,
main_unifiedjobtemplate.last_job_failed,
main_unifiedjobtemplate.last_job_run,
main_unifiedjobtemplate.next_job_run,
main_unifiedjobtemplate.next_schedule_id,
main_unifiedjobtemplate.status
FROM main_unifiedjobtemplate, django_content_type
WHERE main_unifiedjobtemplate.polymorphic_ctype_id = django_content_type.id
ORDER BY main_unifiedjobtemplate.id ASC) TO STDOUT WITH CSV HEADER'''.format(since.strftime("'%Y-%m-%d %H:%M:%S'"))
_copy_table(table='unified_job_template', query=unified_job_template_query, path=full_path)
return

142
awx/main/analytics/core.py Normal file
View File

@@ -0,0 +1,142 @@
import inspect
import json
import logging
import os
import os.path
import tempfile
import shutil
import subprocess
from django.conf import settings
from django.utils.encoding import smart_str
from django.utils.timezone import now, timedelta
from rest_framework.exceptions import PermissionDenied
from awx.conf.license import get_license
from awx.main.models import Job
from awx.main.access import access_registry
from awx.main.models.ha import TowerAnalyticsState
__all__ = ['register', 'gather', 'ship']
logger = logging.getLogger('awx.main.analytics')
def _valid_license():
try:
if get_license(show_key=False).get('license_type', 'UNLICENSED') == 'open':
return False
access_registry[Job](None).check_license()
except PermissionDenied:
logger.exception("A valid license was not found:")
return False
return True
def register(key):
"""
A decorator used to register a function as a metric collector.
Decorated functions should return JSON-serializable objects.
@register('projects_by_scm_type')
def projects_by_scm_type():
return {'git': 5, 'svn': 1, 'hg': 0}
"""
def decorate(f):
f.__awx_analytics_key__ = key
return f
return decorate
def gather(dest=None, module=None):
"""
Gather all defined metrics and write them as JSON files in a .tgz
:param dest: the (optional) absolute path to write a compressed tarball
:pararm module: the module to search for registered analytic collector
functions; defaults to awx.main.analytics.collectors
"""
run_now = now()
state = TowerAnalyticsState.get_solo()
last_run = state.last_run
logger.debug("Last analytics run was: {}".format(last_run))
max_interval = now() - timedelta(days=7)
if last_run < max_interval or not last_run:
last_run = max_interval
if _valid_license() is False:
logger.exception("Invalid License provided, or No License Provided")
return "Error: Invalid License provided, or No License Provided"
if not settings.INSIGHTS_TRACKING_STATE:
logger.error("Insights analytics not enabled")
return
if module is None:
from awx.main.analytics import collectors
module = collectors
dest = dest or tempfile.mkdtemp(prefix='awx_analytics')
for name, func in inspect.getmembers(module):
if inspect.isfunction(func) and hasattr(func, '__awx_analytics_key__'):
key = func.__awx_analytics_key__
path = '{}.json'.format(os.path.join(dest, key))
with open(path, 'w', encoding='utf-8') as f:
try:
json.dump(func(last_run), f)
except Exception:
logger.exception("Could not generate metric {}.json".format(key))
f.close()
os.remove(f.name)
try:
collectors.copy_tables(since=last_run, full_path=dest)
except Exception:
logger.exception("Could not copy tables")
# can't use isoformat() since it has colons, which GNU tar doesn't like
tarname = '_'.join([
settings.SYSTEM_UUID,
run_now.strftime('%Y-%m-%d-%H%M%S%z')
])
tgz = shutil.make_archive(
os.path.join(os.path.dirname(dest), tarname),
'gztar',
dest
)
shutil.rmtree(dest)
return tgz
def ship(path):
"""
Ship gathered metrics via the Insights agent
"""
agent = 'insights-client'
if shutil.which(agent) is None:
logger.error('could not find {} on PATH'.format(agent))
return
logger.debug('shipping analytics file: {}'.format(path))
try:
cmd = [
agent, '--payload', path, '--content-type', settings.INSIGHTS_AGENT_MIME
]
output = smart_str(subprocess.check_output(cmd, timeout=60 * 5))
logger.debug(output)
# reset the `last_run` when data is shipped
run_now = now()
state = TowerAnalyticsState.get_solo()
state.last_run = run_now
state.save()
except subprocess.CalledProcessError:
logger.exception('{} failure:'.format(cmd))
except subprocess.TimeoutExpired:
logger.exception('{} timeout:'.format(cmd))

View File

@@ -0,0 +1,121 @@
from django.conf import settings
from prometheus_client import (
REGISTRY,
PROCESS_COLLECTOR,
PLATFORM_COLLECTOR,
GC_COLLECTOR,
Gauge,
Info,
generate_latest
)
from awx.conf.license import get_license
from awx.main.utils import (get_awx_version, get_ansible_version)
from awx.main.analytics.collectors import (
counts,
instance_info,
job_instance_counts,
)
REGISTRY.unregister(PROCESS_COLLECTOR)
REGISTRY.unregister(PLATFORM_COLLECTOR)
REGISTRY.unregister(GC_COLLECTOR)
SYSTEM_INFO = Info('awx_system', 'AWX System Information')
ORG_COUNT = Gauge('awx_organizations_total', 'Number of organizations')
USER_COUNT = Gauge('awx_users_total', 'Number of users')
TEAM_COUNT = Gauge('awx_teams_total', 'Number of teams')
INV_COUNT = Gauge('awx_inventories_total', 'Number of inventories')
PROJ_COUNT = Gauge('awx_projects_total', 'Number of projects')
JT_COUNT = Gauge('awx_job_templates_total', 'Number of job templates')
WFJT_COUNT = Gauge('awx_workflow_job_templates_total', 'Number of workflow job templates')
HOST_COUNT = Gauge('awx_hosts_total', 'Number of hosts', ['type',])
SCHEDULE_COUNT = Gauge('awx_schedules_total', 'Number of schedules')
INV_SCRIPT_COUNT = Gauge('awx_inventory_scripts_total', 'Number of invetory scripts')
USER_SESSIONS = Gauge('awx_sessions_total', 'Number of sessions', ['type',])
CUSTOM_VENVS = Gauge('awx_custom_virtualenvs_total', 'Number of virtualenvs')
RUNNING_JOBS = Gauge('awx_running_jobs_total', 'Number of running jobs on the Tower system')
INSTANCE_CAPACITY = Gauge('awx_instance_capacity', 'Capacity of each node in a Tower system', ['instance_uuid',])
INSTANCE_CPU = Gauge('awx_instance_cpu', 'CPU cores on each node in a Tower system', ['instance_uuid',])
INSTANCE_MEMORY = Gauge('awx_instance_memory', 'RAM (Kb) on each node in a Tower system', ['instance_uuid',])
INSTANCE_INFO = Info('awx_instance', 'Info about each node in a Tower system', ['instance_uuid',])
INSTANCE_LAUNCH_TYPE = Gauge('awx_instance_launch_type_total', 'Type of Job launched', ['node', 'launch_type',])
INSTANCE_STATUS = Gauge('awx_instance_status_total', 'Status of Job launched', ['node', 'status',])
LICENSE_INSTANCE_TOTAL = Gauge('awx_license_instance_total', 'Total number of managed hosts provided by your license')
LICENSE_INSTANCE_FREE = Gauge('awx_license_instance_free', 'Number of remaining managed hosts provided by your license')
def metrics():
license_info = get_license(show_key=False)
SYSTEM_INFO.info({
'install_uuid': settings.INSTALL_UUID,
'insights_analytics': str(settings.INSIGHTS_TRACKING_STATE),
'tower_url_base': settings.TOWER_URL_BASE,
'tower_version': get_awx_version(),
'ansible_version': get_ansible_version(),
'license_type': license_info.get('license_type', 'UNLICENSED'),
'license_expiry': str(license_info.get('time_remaining', 0)),
'pendo_tracking': settings.PENDO_TRACKING_STATE,
'external_logger_enabled': str(settings.LOG_AGGREGATOR_ENABLED),
'external_logger_type': getattr(settings, 'LOG_AGGREGATOR_TYPE', 'None')
})
LICENSE_INSTANCE_TOTAL.set(str(license_info.get('available_instances', 0)))
LICENSE_INSTANCE_FREE.set(str(license_info.get('free_instances', 0)))
current_counts = counts(None)
ORG_COUNT.set(current_counts['organization'])
USER_COUNT.set(current_counts['user'])
TEAM_COUNT.set(current_counts['team'])
INV_COUNT.set(current_counts['inventory'])
PROJ_COUNT.set(current_counts['project'])
JT_COUNT.set(current_counts['job_template'])
WFJT_COUNT.set(current_counts['workflow_job_template'])
HOST_COUNT.labels(type='all').set(current_counts['host'])
HOST_COUNT.labels(type='active').set(current_counts['active_host_count'])
SCHEDULE_COUNT.set(current_counts['schedule'])
INV_SCRIPT_COUNT.set(current_counts['custom_inventory_script'])
CUSTOM_VENVS.set(current_counts['custom_virtualenvs'])
USER_SESSIONS.labels(type='all').set(current_counts['active_sessions'])
USER_SESSIONS.labels(type='user').set(current_counts['active_user_sessions'])
USER_SESSIONS.labels(type='anonymous').set(current_counts['active_anonymous_sessions'])
RUNNING_JOBS.set(current_counts['running_jobs'])
instance_data = instance_info(None)
for uuid in instance_data:
INSTANCE_CAPACITY.labels(instance_uuid=uuid).set(instance_data[uuid]['capacity'])
INSTANCE_CPU.labels(instance_uuid=uuid).set(instance_data[uuid]['cpu'])
INSTANCE_MEMORY.labels(instance_uuid=uuid).set(instance_data[uuid]['memory'])
INSTANCE_INFO.labels(instance_uuid=uuid).info({
'enabled': str(instance_data[uuid]['enabled']),
'last_isolated_check': getattr(instance_data[uuid], 'last_isolated_check', 'None'),
'managed_by_policy': str(instance_data[uuid]['managed_by_policy']),
'version': instance_data[uuid]['version']
})
instance_data = job_instance_counts(None)
for node in instance_data:
# skipping internal execution node (for system jobs)
if node == '':
continue
types = instance_data[node].get('launch_type', {})
for launch_type, value in types.items():
INSTANCE_LAUNCH_TYPE.labels(node=node, launch_type=launch_type).set(value)
statuses = instance_data[node].get('status', {})
for status, value in statuses.items():
INSTANCE_STATUS.labels(node=node, status=status).set(value)
return generate_latest()
__all__ = ['metrics']

View File

@@ -21,7 +21,6 @@ register(
help_text=_('Enable capturing activity for the activity stream.'),
category=_('System'),
category_slug='system',
feature_required='activity_streams',
)
register(
@@ -31,7 +30,6 @@ register(
help_text=_('Enable capturing activity for the activity stream when running inventory sync.'),
category=_('System'),
category_slug='system',
feature_required='activity_streams',
)
register(
@@ -120,12 +118,32 @@ register(
default=_load_default_license_from_file,
label=_('License'),
help_text=_('The license controls which features and functionality are '
'enabled. Use /api/v1/config/ to update or change '
'enabled. Use /api/v2/config/ to update or change '
'the license.'),
category=_('System'),
category_slug='system',
)
register(
'INSTALL_UUID',
field_class=fields.CharField,
label=_('Unique identifier for an AWX/Tower installation'),
category=_('System'),
category_slug='system',
read_only=True,
)
register(
'CUSTOM_VENV_PATHS',
field_class=fields.StringListPathField,
label=_('Custom virtual environment paths'),
help_text=_('Paths where Tower will look for custom virtual environments '
'(in addition to /var/lib/awx/venv/). Enter one path per line.'),
category=_('System'),
category_slug='system',
default=[],
)
register(
'AD_HOC_COMMANDS',
field_class=fields.StringListField,
@@ -290,6 +308,16 @@ register(
placeholder={'HTTP_PROXY': 'myproxy.local:8080'},
)
register(
'INSIGHTS_TRACKING_STATE',
field_class=fields.BooleanField,
default=False,
label=_('Gather data for Automation Insights'),
help_text=_('Enables Tower to gather data on automation and send it to Red Hat Insights.'),
category=_('System'),
category_slug='system',
)
register(
'AWX_ROLES_ENABLED',
field_class=fields.BooleanField,
@@ -542,6 +570,16 @@ register(
)
register(
'BROKER_DURABILITY',
field_class=fields.BooleanField,
label=_('Message Durability'),
help_text=_('When set (the default), underlying queues will be persisted to disk. Disable this to enable higher message bus throughput.'),
category=_('System'),
category_slug='system',
)
def logging_validate(serializer, attrs):
if not serializer.instance or \
not hasattr(serializer.instance, 'LOG_AGGREGATOR_HOST') or \

View File

@@ -16,7 +16,8 @@ SCHEDULEABLE_PROVIDERS = CLOUD_PROVIDERS + ('custom', 'scm',)
PRIVILEGE_ESCALATION_METHODS = [
('sudo', _('Sudo')), ('su', _('Su')), ('pbrun', _('Pbrun')), ('pfexec', _('Pfexec')),
('dzdo', _('DZDO')), ('pmrun', _('Pmrun')), ('runas', _('Runas')),
('enable', _('Enable')), ('doas', _('Doas')),
('enable', _('Enable')), ('doas', _('Doas')), ('ksu', _('Ksu')),
('machinectl', _('Machinectl')), ('sesu', _('Sesu')),
]
CHOICES_PRIVILEGE_ESCALATION_METHODS = [('', _('None'))] + PRIVILEGE_ESCALATION_METHODS
ANSI_SGR_PATTERN = re.compile(r'\x1b\[[0-9;]*m')
@@ -24,7 +25,9 @@ STANDARD_INVENTORY_UPDATE_ENV = {
# Failure to parse inventory should always be fatal
'ANSIBLE_INVENTORY_UNPARSED_FAILED': 'True',
# Always use the --export option for ansible-inventory
'ANSIBLE_INVENTORY_EXPORT': 'True'
'ANSIBLE_INVENTORY_EXPORT': 'True',
# Redirecting output to stderr allows JSON parsing to still work with -vvv
'ANSIBLE_VERBOSE_TO_STDERR': 'True'
}
CAN_CANCEL = ('new', 'pending', 'waiting', 'running')
ACTIVE_STATES = CAN_CANCEL
@@ -34,6 +37,17 @@ ENV_BLACKLIST = frozenset((
'INVENTORY_ID', 'INVENTORY_SOURCE_ID', 'INVENTORY_UPDATE_ID',
'AD_HOC_COMMAND_ID', 'REST_API_URL', 'REST_API_TOKEN', 'MAX_EVENT_RES',
'CALLBACK_QUEUE', 'CALLBACK_CONNECTION', 'CACHE',
'JOB_CALLBACK_DEBUG', 'INVENTORY_HOSTVARS', 'FACT_QUEUE',
'JOB_CALLBACK_DEBUG', 'INVENTORY_HOSTVARS',
'AWX_HOST', 'PROJECT_REVISION'
))
# loggers that may be called in process of emitting a log
LOGGER_BLACKLIST = (
'awx.main.utils.handlers',
'awx.main.utils.formatters',
'awx.main.utils.filters',
'awx.main.utils.encryption',
'awx.main.utils.log',
# loggers that may be called getting logging settings
'awx.conf'
)

View File

@@ -0,0 +1,125 @@
from .plugin import CredentialPlugin
import os
import stat
import tempfile
import threading
from urllib.parse import quote, urlencode, urljoin
from django.utils.translation import ugettext_lazy as _
import requests
aim_inputs = {
'fields': [{
'id': 'url',
'label': _('CyberArk AIM URL'),
'type': 'string',
'format': 'url',
}, {
'id': 'app_id',
'label': _('Application ID'),
'type': 'string',
'secret': True,
}, {
'id': 'client_key',
'label': _('Client Key'),
'type': 'string',
'secret': True,
'multiline': True,
}, {
'id': 'client_cert',
'label': _('Client Certificate'),
'type': 'string',
'secret': True,
'multiline': True,
}, {
'id': 'verify',
'label': _('Verify SSL Certificates'),
'type': 'boolean',
'default': True,
}],
'metadata': [{
'id': 'object_query',
'label': _('Object Query'),
'type': 'string',
'help_text': _('Lookup query for the object. Ex: "Safe=TestSafe;Object=testAccountName123"'),
}, {
'id': 'object_query_format',
'label': _('Object Query Format'),
'type': 'string',
'default': 'Exact',
'choices': ['Exact', 'Regexp']
}, {
'id': 'reason',
'label': _('Reason'),
'type': 'string',
'help_text': _('Object request reason. This is only needed if it is required by the object\'s policy.')
}],
'required': ['url', 'app_id', 'object_query'],
}
def create_temporary_fifo(data):
"""Open fifo named pipe in a new thread using a temporary file path. The
thread blocks until data is read from the pipe.
Returns the path to the fifo.
:param data(bytes): Data to write to the pipe.
"""
path = os.path.join(tempfile.mkdtemp(), next(tempfile._get_candidate_names()))
os.mkfifo(path, stat.S_IRUSR | stat.S_IWUSR)
threading.Thread(
target=lambda p, d: open(p, 'wb').write(d),
args=(path, data)
).start()
return path
def aim_backend(**kwargs):
url = kwargs['url']
client_cert = kwargs.get('client_cert', None)
client_key = kwargs.get('client_key', None)
verify = kwargs['verify']
app_id = kwargs['app_id']
object_query = kwargs['object_query']
object_query_format = kwargs['object_query_format']
reason = kwargs.get('reason', None)
query_params = {
'AppId': app_id,
'Query': object_query,
'QueryFormat': object_query_format,
}
if reason:
query_params['reason'] = reason
request_qs = '?' + urlencode(query_params, quote_via=quote)
request_url = urljoin(url, '/'.join(['AIMWebService', 'api', 'Accounts']))
cert = None
if client_cert and client_key:
cert = (
create_temporary_fifo(client_cert.encode()),
create_temporary_fifo(client_key.encode())
)
elif client_cert:
cert = create_temporary_fifo(client_cert.encode())
res = requests.get(
request_url + request_qs,
timeout=30,
cert=cert,
verify=verify,
)
res.raise_for_status()
return res.json()['Content']
aim_plugin = CredentialPlugin(
'CyberArk AIM Secret Lookup',
inputs=aim_inputs,
backend=aim_backend
)

View File

@@ -0,0 +1,65 @@
from .plugin import CredentialPlugin
from django.utils.translation import ugettext_lazy as _
from azure.keyvault import KeyVaultClient, KeyVaultAuthentication
from azure.common.credentials import ServicePrincipalCredentials
azure_keyvault_inputs = {
'fields': [{
'id': 'url',
'label': _('Vault URL (DNS Name)'),
'type': 'string',
'format': 'url',
}, {
'id': 'client',
'label': _('Client ID'),
'type': 'string'
}, {
'id': 'secret',
'label': _('Client Secret'),
'type': 'string',
'secret': True,
}, {
'id': 'tenant',
'label': _('Tenant ID'),
'type': 'string'
}],
'metadata': [{
'id': 'secret_field',
'label': _('Secret Name'),
'type': 'string',
'help_text': _('The name of the secret to look up.'),
}, {
'id': 'secret_version',
'label': _('Secret Version'),
'type': 'string',
'help_text': _('Used to specify a specific secret version (if left empty, the latest version will be used).'),
}],
'required': ['url', 'client', 'secret', 'tenant', 'secret_field'],
}
def azure_keyvault_backend(**kwargs):
url = kwargs['url']
def auth_callback(server, resource, scope):
credentials = ServicePrincipalCredentials(
url = url,
client_id = kwargs['client'],
secret = kwargs['secret'],
tenant = kwargs['tenant'],
resource = "https://vault.azure.net",
)
token = credentials.token
return token['token_type'], token['access_token']
kv = KeyVaultClient(KeyVaultAuthentication(auth_callback))
return kv.get_secret(url, kwargs['secret_field'], kwargs.get('secret_version', '')).value
azure_keyvault_plugin = CredentialPlugin(
'Microsoft Azure Key Vault',
inputs=azure_keyvault_inputs,
backend=azure_keyvault_backend
)

View File

@@ -0,0 +1,121 @@
from .plugin import CredentialPlugin
import base64
import os
import stat
import tempfile
import threading
from urllib.parse import urljoin, quote_plus
from django.utils.translation import ugettext_lazy as _
import requests
conjur_inputs = {
'fields': [{
'id': 'url',
'label': _('Conjur URL'),
'type': 'string',
'format': 'url',
}, {
'id': 'api_key',
'label': _('API Key'),
'type': 'string',
'secret': True,
}, {
'id': 'account',
'label': _('Account'),
'type': 'string',
}, {
'id': 'username',
'label': _('Username'),
'type': 'string',
}, {
'id': 'cacert',
'label': _('Public Key Certificate'),
'type': 'string',
'multiline': True
}],
'metadata': [{
'id': 'secret_path',
'label': _('Secret Identifier'),
'type': 'string',
'help_text': _('The identifier for the secret e.g., /some/identifier'),
}, {
'id': 'secret_version',
'label': _('Secret Version'),
'type': 'string',
'help_text': _('Used to specify a specific secret version (if left empty, the latest version will be used).'),
}],
'required': ['url', 'api_key', 'account', 'username'],
}
def create_temporary_fifo(data):
"""Open fifo named pipe in a new thread using a temporary file path. The
thread blocks until data is read from the pipe.
Returns the path to the fifo.
:param data(bytes): Data to write to the pipe.
"""
path = os.path.join(tempfile.mkdtemp(), next(tempfile._get_candidate_names()))
os.mkfifo(path, stat.S_IRUSR | stat.S_IWUSR)
threading.Thread(
target=lambda p, d: open(p, 'wb').write(d),
args=(path, data)
).start()
return path
def conjur_backend(**kwargs):
url = kwargs['url']
api_key = kwargs['api_key']
account = quote_plus(kwargs['account'])
username = quote_plus(kwargs['username'])
secret_path = quote_plus(kwargs['secret_path'])
version = kwargs.get('secret_version')
cacert = kwargs.get('cacert', None)
auth_kwargs = {
'headers': {'Content-Type': 'text/plain'},
'data': api_key
}
if cacert:
auth_kwargs['verify'] = create_temporary_fifo(cacert.encode())
# https://www.conjur.org/api.html#authentication-authenticate-post
resp = requests.post(
urljoin(url, '/'.join(['authn', account, username, 'authenticate'])),
**auth_kwargs
)
resp.raise_for_status()
token = base64.b64encode(resp.content).decode('utf-8')
lookup_kwargs = {
'headers': {'Authorization': 'Token token="{}"'.format(token)},
}
if cacert:
lookup_kwargs['verify'] = create_temporary_fifo(cacert.encode())
# https://www.conjur.org/api.html#secrets-retrieve-a-secret-get
path = urljoin(url, '/'.join([
'secrets',
account,
'variable',
secret_path
]))
if version:
path = '?'.join([path, version])
resp = requests.get(path, timeout=30, **lookup_kwargs)
resp.raise_for_status()
return resp.text
conjur_plugin = CredentialPlugin(
'CyberArk Conjur Secret Lookup',
inputs=conjur_inputs,
backend=conjur_backend
)

View File

@@ -0,0 +1,152 @@
import copy
import os
import pathlib
from urllib.parse import urljoin
from .plugin import CredentialPlugin
import requests
from django.utils.translation import ugettext_lazy as _
base_inputs = {
'fields': [{
'id': 'url',
'label': _('Server URL'),
'type': 'string',
'format': 'url',
'help_text': _('The URL to the HashiCorp Vault'),
}, {
'id': 'token',
'label': _('Token'),
'type': 'string',
'secret': True,
'help_text': _('The access token used to authenticate to the Vault server'),
}],
'metadata': [{
'id': 'secret_path',
'label': _('Path to Secret'),
'type': 'string',
'help_text': _('The path to the secret e.g., /some-engine/some-secret/'),
}],
'required': ['url', 'token', 'secret_path'],
}
hashi_kv_inputs = copy.deepcopy(base_inputs)
hashi_kv_inputs['fields'].append({
'id': 'api_version',
'label': _('API Version'),
'choices': ['v1', 'v2'],
'help_text': _('API v1 is for static key/value lookups. API v2 is for versioned key/value lookups.'),
'default': 'v1',
})
hashi_kv_inputs['metadata'].extend([{
'id': 'secret_key',
'label': _('Key Name'),
'type': 'string',
'help_text': _('The name of the key to look up in the secret.'),
}, {
'id': 'secret_version',
'label': _('Secret Version (v2 only)'),
'type': 'string',
'help_text': _('Used to specify a specific secret version (if left empty, the latest version will be used).'),
}])
hashi_kv_inputs['required'].extend(['api_version', 'secret_key'])
hashi_ssh_inputs = copy.deepcopy(base_inputs)
hashi_ssh_inputs['metadata'] = [{
'id': 'public_key',
'label': _('Unsigned Public Key'),
'type': 'string',
'multiline': True,
}] + hashi_ssh_inputs['metadata'] + [{
'id': 'role',
'label': _('Role Name'),
'type': 'string',
'help_text': _('The name of the role used to sign.')
}, {
'id': 'valid_principals',
'label': _('Valid Principals'),
'type': 'string',
'help_text': _('Valid principals (either usernames or hostnames) that the certificate should be signed for.'),
}]
hashi_ssh_inputs['required'].extend(['public_key', 'role'])
def kv_backend(**kwargs):
token = kwargs['token']
url = urljoin(kwargs['url'], 'v1')
secret_path = kwargs['secret_path']
secret_key = kwargs.get('secret_key', None)
api_version = kwargs['api_version']
sess = requests.Session()
sess.headers['Authorization'] = 'Bearer {}'.format(token)
if api_version == 'v2':
params = {}
if kwargs.get('secret_version'):
params['version'] = kwargs['secret_version']
try:
mount_point, *path = pathlib.Path(secret_path.lstrip(os.sep)).parts
'/'.join(*path)
except Exception:
mount_point, path = secret_path, []
# https://www.vaultproject.io/api/secret/kv/kv-v2.html#read-secret-version
response = sess.get(
'/'.join([url, mount_point, 'data'] + path).rstrip('/'),
params=params,
timeout=30
)
response.raise_for_status()
json = response.json()['data']
else:
# https://www.vaultproject.io/api/secret/kv/kv-v1.html#read-secret
response = sess.get('/'.join([url, secret_path]).rstrip('/'), timeout=30)
response.raise_for_status()
json = response.json()
if secret_key:
try:
return json['data'][secret_key]
except KeyError:
raise RuntimeError(
'{} is not present at {}'.format(secret_key, secret_path)
)
return json['data']
def ssh_backend(**kwargs):
token = kwargs['token']
url = urljoin(kwargs['url'], 'v1')
secret_path = kwargs['secret_path']
role = kwargs['role']
sess = requests.Session()
sess.headers['Authorization'] = 'Bearer {}'.format(token)
json = {
'public_key': kwargs['public_key']
}
if kwargs.get('valid_principals'):
json['valid_principals'] = kwargs['valid_principals']
# https://www.vaultproject.io/api/secret/ssh/index.html#sign-ssh-key
resp = sess.post(
'/'.join([url, secret_path, 'sign', role]).rstrip('/'),
json=json,
timeout=30
)
resp.raise_for_status()
return resp.json()['data']['signed_key']
hashivault_kv_plugin = CredentialPlugin(
'HashiCorp Vault Secret Lookup',
inputs=hashi_kv_inputs,
backend=kv_backend
)
hashivault_ssh_plugin = CredentialPlugin(
'HashiCorp Vault Signed SSH',
inputs=hashi_ssh_inputs,
backend=ssh_backend
)

View File

@@ -0,0 +1,3 @@
from collections import namedtuple
CredentialPlugin = namedtuple('CredentialPlugin', ['name', 'inputs', 'backend'])

View File

@@ -0,0 +1,155 @@
import os
import pkg_resources
import sqlite3
import sys
import traceback
import uuid
from django.core.cache import cache
from django.core.cache.backends.locmem import LocMemCache
from django.db.backends.postgresql.base import DatabaseWrapper as BaseDatabaseWrapper
from awx.main.utils import memoize
__loc__ = LocMemCache(str(uuid.uuid4()), {})
__all__ = ['DatabaseWrapper']
class RecordedQueryLog(object):
def __init__(self, log, db, dest='/var/log/tower/profile'):
self.log = log
self.db = db
self.dest = dest
try:
self.threshold = cache.get('awx-profile-sql-threshold')
except Exception:
# if we can't reach memcached, just assume profiling's off
self.threshold = None
def append(self, query):
ret = self.log.append(query)
try:
self.write(query)
except Exception:
# not sure what else to do her e- we can't really safely
# *use* our loggers because it'll just generate more DB queries
# and potentially recurse into this state again
_, _, tb = sys.exc_info()
traceback.print_tb(tb)
return ret
def write(self, query):
if self.threshold is None:
return
seconds = float(query['time'])
# if the query is slow enough...
if seconds >= self.threshold:
sql = query['sql']
if sql.startswith('EXPLAIN'):
return
# build a printable Python stack
bt = ' '.join(traceback.format_stack())
# and re-run the same query w/ EXPLAIN
explain = ''
cursor = self.db.cursor()
cursor.execute('EXPLAIN VERBOSE {}'.format(sql))
for line in cursor.fetchall():
explain += line[0] + '\n'
# write a row of data into a per-PID sqlite database
if not os.path.isdir(self.dest):
os.makedirs(self.dest)
progname = ' '.join(sys.argv)
for match in ('uwsgi', 'dispatcher', 'callback_receiver', 'runworker'):
if match in progname:
progname = match
break
else:
progname = os.path.basename(sys.argv[0])
filepath = os.path.join(
self.dest,
'{}.sqlite'.format(progname)
)
version = pkg_resources.get_distribution('awx').version
log = sqlite3.connect(filepath, timeout=3)
log.execute(
'CREATE TABLE IF NOT EXISTS queries ('
' id INTEGER PRIMARY KEY,'
' version TEXT,'
' pid INTEGER,'
' stamp DATETIME DEFAULT CURRENT_TIMESTAMP,'
' argv REAL,'
' time REAL,'
' sql TEXT,'
' explain TEXT,'
' bt TEXT'
');'
)
log.commit()
log.execute(
'INSERT INTO queries (pid, version, argv, time, sql, explain, bt) '
'VALUES (?, ?, ?, ?, ?, ?, ?);',
(os.getpid(), version, ' ' .join(sys.argv), seconds, sql, explain, bt)
)
log.commit()
def __len__(self):
return len(self.log)
def __iter__(self):
return iter(self.log)
def __getattr__(self, attr):
return getattr(self.log, attr)
class DatabaseWrapper(BaseDatabaseWrapper):
"""
This is a special subclass of Django's postgres DB backend which - based on
the value of a special flag in memcached - captures slow queries and
writes profile and Python stack metadata to the disk.
"""
def __init__(self, *args, **kwargs):
super(DatabaseWrapper, self).__init__(*args, **kwargs)
# Django's default base wrapper implementation has `queries_log`
# which is a `collections.deque` that every query is appended to
#
# this line wraps the deque with a proxy that can capture each query
# and - if it's slow enough - record profiling metadata to the file
# system for debugging purposes
self.queries_log = RecordedQueryLog(self.queries_log, self)
@property
@memoize(ttl=1, cache=__loc__)
def force_debug_cursor(self):
# in Django's base DB implementation, `self.force_debug_cursor` is just
# a simple boolean, and this value is used to signal to Django that it
# should record queries into `self.queries_log` as they're executed (this
# is the same mechanism used by libraries like the django-debug-toolbar)
#
# in _this_ implementation, we represent it as a property which will
# check memcache for a special flag to be set (when the flag is set, it
# means we should start recording queries because somebody called
# `awx-manage profile_sql`)
#
# it's worth noting that this property is wrapped w/ @memoize because
# Django references this attribute _constantly_ (in particular, once
# per executed query); doing a memcached.get() _at most_ once per
# second is a good enough window to detect when profiling is turned
# on/off by a system administrator
try:
threshold = cache.get('awx-profile-sql-threshold')
except Exception:
# if we can't reach memcached, just assume profiling's off
threshold = None
self.queries_log.threshold = threshold
return threshold is not None
@force_debug_cursor.setter
def force_debug_cursor(self, v):
return

View File

@@ -4,7 +4,8 @@ import socket
from django.conf import settings
from awx.main.dispatch import get_local_queuename
from kombu import Connection, Queue, Exchange, Producer, Consumer
from awx.main.dispatch.kombu import Connection
from kombu import Queue, Exchange, Producer, Consumer
logger = logging.getLogger('awx.main.dispatch')

View File

@@ -0,0 +1,42 @@
from amqp.exceptions import PreconditionFailed
from django.conf import settings
from kombu.connection import Connection as KombuConnection
from kombu.transport import pyamqp
import logging
logger = logging.getLogger('awx.main.dispatch')
__all__ = ['Connection']
class Connection(KombuConnection):
def __init__(self, *args, **kwargs):
super(Connection, self).__init__(*args, **kwargs)
class _Channel(pyamqp.Channel):
def queue_declare(self, queue, *args, **kwargs):
kwargs['durable'] = settings.BROKER_DURABILITY
try:
return super(_Channel, self).queue_declare(queue, *args, **kwargs)
except PreconditionFailed as e:
if "inequivalent arg 'durable'" in getattr(e, 'reply_text', None):
logger.error(
'queue {} durability is not {}, deleting and recreating'.format(
queue,
kwargs['durable']
)
)
self.queue_delete(queue)
return super(_Channel, self).queue_declare(queue, *args, **kwargs)
class _Connection(pyamqp.Connection):
Channel = _Channel
class _Transport(pyamqp.Transport):
Connection = _Connection
self.transport_cls = _Transport

View File

@@ -1,6 +1,7 @@
import logging
import os
import random
import sys
import traceback
from uuid import uuid4
@@ -18,7 +19,10 @@ import psutil
from awx.main.models import UnifiedJob
from awx.main.dispatch import reaper
logger = logging.getLogger('awx.main.dispatch')
if 'run_callback_receiver' in sys.argv:
logger = logging.getLogger('awx.main.commands.run_callback_receiver')
else:
logger = logging.getLogger('awx.main.dispatch')
class PoolWorker(object):

View File

@@ -4,7 +4,9 @@ import sys
from uuid import uuid4
from django.conf import settings
from kombu import Connection, Exchange, Producer
from kombu import Exchange, Producer
from awx.main.dispatch.kombu import Connection
logger = logging.getLogger('awx.main.dispatch')

View File

@@ -4,6 +4,7 @@
import os
import logging
import signal
import sys
from uuid import UUID
from queue import Empty as QueueEmpty
@@ -13,7 +14,10 @@ from kombu.mixins import ConsumerMixin
from awx.main.dispatch.pool import WorkerPool
logger = logging.getLogger('awx.main.dispatch')
if 'run_callback_receiver' in sys.argv:
logger = logging.getLogger('awx.main.commands.run_callback_receiver')
else:
logger = logging.getLogger('awx.main.dispatch')
def signame(sig):
@@ -108,7 +112,7 @@ class AWXConsumer(ConsumerMixin):
def stop(self, signum, frame):
self.should_stop = True # this makes the kombu mixin stop consuming
logger.debug('received {}, stopping'.format(signame(signum)))
logger.warn('received {}, stopping'.format(signame(signum)))
self.worker.on_stop()
raise SystemExit()

View File

@@ -20,8 +20,8 @@ class CallbackBrokerWorker(BaseWorker):
A worker implementation that deserializes callback event data and persists
it into the database.
The code that *builds* these types of messages is found in the AWX display
callback (`awx.lib.awx_display_callback`).
The code that *generates* these types of messages is found in the
ansible-runner display callback plugin.
'''
MAX_RETRIES = 2

View File

@@ -1,507 +0,0 @@
import base64
import codecs
import json
import os
import shutil
import stat
import tempfile
import time
import logging
from distutils.version import LooseVersion as Version
from io import StringIO
from django.conf import settings
from django.utils.encoding import smart_bytes, smart_str
import awx
from awx.main.expect import run
from awx.main.utils import OutputEventFilter, get_system_task_capacity
from awx.main.queue import CallbackQueueDispatcher
logger = logging.getLogger('awx.isolated.manager')
playbook_logger = logging.getLogger('awx.isolated.manager.playbooks')
class IsolatedManager(object):
def __init__(self, args, cwd, env, stdout_handle, ssh_key_path,
expect_passwords={}, cancelled_callback=None, job_timeout=0,
idle_timeout=None, extra_update_fields=None,
pexpect_timeout=5, proot_cmd='bwrap'):
"""
:param args: a list of `subprocess.call`-style arguments
representing a subprocess e.g.,
['ansible-playbook', '...']
:param cwd: the directory where the subprocess should run,
generally the directory where playbooks exist
:param env: a dict containing environment variables for the
subprocess, ala `os.environ`
:param stdout_handle: a file-like object for capturing stdout
:param ssh_key_path: a filepath where SSH key data can be read
:param expect_passwords: a dict of regular expression password prompts
to input values, i.e., {r'Password:*?$':
'some_password'}
:param cancelled_callback: a callable - which returns `True` or `False`
- signifying if the job has been prematurely
cancelled
:param job_timeout a timeout (in seconds); if the total job runtime
exceeds this, the process will be killed
:param idle_timeout a timeout (in seconds); if new output is not
sent to stdout in this interval, the process
will be terminated
:param extra_update_fields: a dict used to specify DB fields which should
be updated on the underlying model
object after execution completes
:param pexpect_timeout a timeout (in seconds) to wait on
`pexpect.spawn().expect()` calls
:param proot_cmd the command used to isolate processes, `bwrap`
"""
self.args = args
self.cwd = cwd
self.isolated_env = self._redact_isolated_env(env.copy())
self.management_env = self._base_management_env()
self.stdout_handle = stdout_handle
self.ssh_key_path = ssh_key_path
self.expect_passwords = {k.pattern: v for k, v in expect_passwords.items()}
self.cancelled_callback = cancelled_callback
self.job_timeout = job_timeout
self.idle_timeout = idle_timeout
self.extra_update_fields = extra_update_fields
self.pexpect_timeout = pexpect_timeout
self.proot_cmd = proot_cmd
self.started_at = None
@staticmethod
def _base_management_env():
'''
Returns environment variables to use when running a playbook
that manages the isolated instance.
Use of normal job callback and other such configurations are avoided.
'''
env = dict(os.environ.items())
env['ANSIBLE_RETRY_FILES_ENABLED'] = 'False'
env['ANSIBLE_HOST_KEY_CHECKING'] = 'False'
env['ANSIBLE_LIBRARY'] = os.path.join(os.path.dirname(awx.__file__), 'plugins', 'isolated')
return env
@staticmethod
def _build_args(playbook, hosts, extra_vars=None):
'''
Returns list of Ansible CLI command arguments for a management task
:param playbook: name of the playbook to run
:param hosts: host pattern to operate on, ex. "localhost,"
:param extra_vars: optional dictionary of extra_vars to apply
'''
args = [
'ansible-playbook',
playbook,
'-u', settings.AWX_ISOLATED_USERNAME,
'-T', str(settings.AWX_ISOLATED_CONNECTION_TIMEOUT),
'-i', hosts
]
if extra_vars:
args.extend(['-e', json.dumps(extra_vars)])
if settings.AWX_ISOLATED_VERBOSITY:
args.append('-%s' % ('v' * min(5, settings.AWX_ISOLATED_VERBOSITY)))
return args
@staticmethod
def _redact_isolated_env(env):
'''
strips some environment variables that aren't applicable to
job execution within the isolated instance
'''
for var in (
'HOME', 'RABBITMQ_HOST', 'RABBITMQ_PASS', 'RABBITMQ_USER', 'CACHE',
'DJANGO_PROJECT_DIR', 'DJANGO_SETTINGS_MODULE', 'RABBITMQ_VHOST'):
env.pop(var, None)
return env
@classmethod
def awx_playbook_path(cls):
return os.path.abspath(os.path.join(
os.path.dirname(awx.__file__),
'playbooks'
))
def path_to(self, *args):
return os.path.join(self.private_data_dir, *args)
def dispatch(self):
'''
Compile the playbook, its environment, and metadata into a series
of files, and ship to a remote host for isolated execution.
'''
self.started_at = time.time()
secrets = {
'env': self.isolated_env,
'passwords': self.expect_passwords,
'ssh_key_data': None,
'idle_timeout': self.idle_timeout,
'job_timeout': self.job_timeout,
'pexpect_timeout': self.pexpect_timeout
}
# if an ssh private key fifo exists, read its contents and delete it
if self.ssh_key_path:
buff = StringIO()
with open(self.ssh_key_path, 'r') as fifo:
for line in fifo:
buff.write(line)
secrets['ssh_key_data'] = buff.getvalue()
os.remove(self.ssh_key_path)
# write the entire secret payload to a named pipe
# the run_isolated.yml playbook will use a lookup to read this data
# into a variable, and will replicate the data into a named pipe on the
# isolated instance
secrets_path = os.path.join(self.private_data_dir, 'env')
run.open_fifo_write(
secrets_path,
smart_str(base64.b64encode(smart_bytes(json.dumps(secrets))))
)
self.build_isolated_job_data()
extra_vars = {
'src': self.private_data_dir,
'dest': settings.AWX_PROOT_BASE_PATH,
}
if self.proot_temp_dir:
extra_vars['proot_temp_dir'] = self.proot_temp_dir
# Run ansible-playbook to launch a job on the isolated host. This:
#
# - sets up a temporary directory for proot/bwrap (if necessary)
# - copies encrypted job data from the controlling host to the isolated host (with rsync)
# - writes the encryption secret to a named pipe on the isolated host
# - launches the isolated playbook runner via `awx-expect start <job-id>`
args = self._build_args('run_isolated.yml', '%s,' % self.host, extra_vars)
if self.instance.verbosity:
args.append('-%s' % ('v' * min(5, self.instance.verbosity)))
buff = StringIO()
logger.debug('Starting job {} on isolated host with `run_isolated.yml` playbook.'.format(self.instance.id))
status, rc = IsolatedManager.run_pexpect(
args, self.awx_playbook_path(), self.management_env, buff,
idle_timeout=self.idle_timeout,
job_timeout=settings.AWX_ISOLATED_LAUNCH_TIMEOUT,
pexpect_timeout=5
)
output = buff.getvalue().encode('utf-8')
playbook_logger.info('Isolated job {} dispatch:\n{}'.format(self.instance.id, output))
if status != 'successful':
self.stdout_handle.write(output)
return status, rc
@classmethod
def run_pexpect(cls, pexpect_args, *args, **kw):
isolated_ssh_path = None
try:
if all([
getattr(settings, 'AWX_ISOLATED_KEY_GENERATION', False) is True,
getattr(settings, 'AWX_ISOLATED_PRIVATE_KEY', None)
]):
isolated_ssh_path = tempfile.mkdtemp(prefix='awx_isolated', dir=settings.AWX_PROOT_BASE_PATH)
os.chmod(isolated_ssh_path, stat.S_IRUSR | stat.S_IWUSR | stat.S_IXUSR)
isolated_key = os.path.join(isolated_ssh_path, '.isolated')
ssh_sock = os.path.join(isolated_ssh_path, '.isolated_ssh_auth.sock')
run.open_fifo_write(isolated_key, settings.AWX_ISOLATED_PRIVATE_KEY)
pexpect_args = run.wrap_args_with_ssh_agent(pexpect_args, isolated_key, ssh_sock, silence_ssh_add=True)
return run.run_pexpect(pexpect_args, *args, **kw)
finally:
if isolated_ssh_path:
shutil.rmtree(isolated_ssh_path)
def build_isolated_job_data(self):
'''
Write the playbook and metadata into a collection of files on the local
file system.
This function is intended to be used to compile job data so that it
can be shipped to a remote, isolated host (via ssh).
'''
rsync_exclude = [
# don't rsync source control metadata (it can be huge!)
'- /project/.git',
'- /project/.svn',
'- /project/.hg',
# don't rsync job events that are in the process of being written
'- /artifacts/job_events/*-partial.json.tmp',
# rsync can't copy named pipe data - we're replicating this manually ourselves in the playbook
'- /env'
]
for filename, data in (
['.rsync-filter', '\n'.join(rsync_exclude)],
['args', json.dumps(self.args)]
):
path = self.path_to(filename)
with open(path, 'w') as f:
f.write(data)
os.chmod(path, stat.S_IRUSR)
# symlink the scm checkout (if there is one) so that it's rsync'ed over, too
if 'AD_HOC_COMMAND_ID' not in self.isolated_env:
os.symlink(self.cwd, self.path_to('project'))
# create directories for build artifacts to live in
os.makedirs(self.path_to('artifacts', 'job_events'), mode=stat.S_IXUSR + stat.S_IWUSR + stat.S_IRUSR)
def _missing_artifacts(self, path_list, output):
missing_artifacts = list(filter(lambda path: not os.path.exists(path), path_list))
for path in missing_artifacts:
self.stdout_handle.write('ansible did not exit cleanly, missing `{}`.\n'.format(path))
if missing_artifacts:
daemon_path = self.path_to('artifacts', 'daemon.log')
if os.path.exists(daemon_path):
# If available, show log files from the run.py call
with codecs.open(daemon_path, 'r', encoding='utf-8') as f:
self.stdout_handle.write(f.read())
else:
# Provide the management playbook standard out if not available
self.stdout_handle.write(output)
return True
return False
def check(self, interval=None):
"""
Repeatedly poll the isolated node to determine if the job has run.
On success, copy job artifacts to the controlling node.
On failure, continue to poll the isolated node (until the job timeout
is exceeded).
For a completed job run, this function returns (status, rc),
representing the status and return code of the isolated
`ansible-playbook` run.
:param interval: an interval (in seconds) to wait between status polls
"""
interval = interval if interval is not None else settings.AWX_ISOLATED_CHECK_INTERVAL
extra_vars = {'src': self.private_data_dir}
args = self._build_args('check_isolated.yml', '%s,' % self.host, extra_vars)
if self.instance.verbosity:
args.append('-%s' % ('v' * min(5, self.instance.verbosity)))
status = 'failed'
output = ''
rc = None
buff = StringIO()
last_check = time.time()
seek = 0
job_timeout = remaining = self.job_timeout
while status == 'failed':
if job_timeout != 0:
remaining = max(0, job_timeout - (time.time() - self.started_at))
if remaining == 0:
# if it takes longer than $REMAINING_JOB_TIMEOUT to retrieve
# job artifacts from the host, consider the job failed
if isinstance(self.extra_update_fields, dict):
self.extra_update_fields['job_explanation'] = "Job terminated due to timeout"
status = 'failed'
break
canceled = self.cancelled_callback() if self.cancelled_callback else False
if not canceled and time.time() - last_check < interval:
# If the job isn't cancelled, but we haven't waited `interval` seconds, wait longer
time.sleep(1)
continue
buff = StringIO()
logger.debug('Checking on isolated job {} with `check_isolated.yml`.'.format(self.instance.id))
status, rc = IsolatedManager.run_pexpect(
args, self.awx_playbook_path(), self.management_env, buff,
cancelled_callback=self.cancelled_callback,
idle_timeout=remaining,
job_timeout=remaining,
pexpect_timeout=5,
proot_cmd=self.proot_cmd
)
output = buff.getvalue().encode('utf-8')
playbook_logger.info('Isolated job {} check:\n{}'.format(self.instance.id, output))
path = self.path_to('artifacts', 'stdout')
if os.path.exists(path):
with codecs.open(path, 'r', encoding='utf-8') as f:
f.seek(seek)
for line in f:
self.stdout_handle.write(line)
seek += len(line)
last_check = time.time()
if status == 'successful':
status_path = self.path_to('artifacts', 'status')
rc_path = self.path_to('artifacts', 'rc')
if self._missing_artifacts([status_path, rc_path], output):
status = 'failed'
rc = 1
else:
with open(status_path, 'r') as f:
status = f.readline()
with open(rc_path, 'r') as f:
rc = int(f.readline())
elif status == 'failed':
# if we were unable to retrieve job reults from the isolated host,
# print stdout of the `check_isolated.yml` playbook for clues
self.stdout_handle.write(smart_str(output))
return status, rc
def cleanup(self):
# If the job failed for any reason, make a last-ditch effort at cleanup
extra_vars = {
'private_data_dir': self.private_data_dir,
'cleanup_dirs': [
self.private_data_dir,
self.proot_temp_dir,
],
}
args = self._build_args('clean_isolated.yml', '%s,' % self.host, extra_vars)
logger.debug('Cleaning up job {} on isolated host with `clean_isolated.yml` playbook.'.format(self.instance.id))
buff = StringIO()
timeout = max(60, 2 * settings.AWX_ISOLATED_CONNECTION_TIMEOUT)
status, rc = IsolatedManager.run_pexpect(
args, self.awx_playbook_path(), self.management_env, buff,
idle_timeout=timeout, job_timeout=timeout,
pexpect_timeout=5
)
output = buff.getvalue().encode('utf-8')
playbook_logger.info('Isolated job {} cleanup:\n{}'.format(self.instance.id, output))
if status != 'successful':
# stdout_handle is closed by this point so writing output to logs is our only option
logger.warning('Isolated job {} cleanup error, output:\n{}'.format(self.instance.id, output))
@classmethod
def update_capacity(cls, instance, task_result, awx_application_version):
instance.version = task_result['version']
isolated_version = instance.version.split("-", 1)[0]
cluster_version = awx_application_version.split("-", 1)[0]
if Version(cluster_version) > Version(isolated_version):
err_template = "Isolated instance {} reports version {}, cluster node is at {}, setting capacity to zero."
logger.error(err_template.format(instance.hostname, instance.version, awx_application_version))
instance.capacity = 0
else:
if instance.capacity == 0 and task_result['capacity_cpu']:
logger.warning('Isolated instance {} has re-joined.'.format(instance.hostname))
instance.cpu_capacity = int(task_result['capacity_cpu'])
instance.mem_capacity = int(task_result['capacity_mem'])
instance.capacity = get_system_task_capacity(scale=instance.capacity_adjustment,
cpu_capacity=int(task_result['capacity_cpu']),
mem_capacity=int(task_result['capacity_mem']))
instance.save(update_fields=['cpu_capacity', 'mem_capacity', 'capacity', 'version', 'modified'])
@classmethod
def health_check(cls, instance_qs, awx_application_version):
'''
:param instance_qs: List of Django objects representing the
isolated instances to manage
Runs playbook that will
- determine if instance is reachable
- find the instance capacity
- clean up orphaned private files
Performs save on each instance to update its capacity.
'''
hostname_string = ''
for instance in instance_qs:
hostname_string += '{},'.format(instance.hostname)
args = cls._build_args('heartbeat_isolated.yml', hostname_string)
args.extend(['--forks', str(len(instance_qs))])
env = cls._base_management_env()
try:
facts_path = tempfile.mkdtemp()
env['ANSIBLE_CACHE_PLUGIN'] = 'jsonfile'
env['ANSIBLE_CACHE_PLUGIN_CONNECTION'] = facts_path
buff = StringIO()
timeout = max(60, 2 * settings.AWX_ISOLATED_CONNECTION_TIMEOUT)
status, rc = IsolatedManager.run_pexpect(
args, cls.awx_playbook_path(), env, buff,
idle_timeout=timeout, job_timeout=timeout,
pexpect_timeout=5
)
heartbeat_stdout = buff.getvalue().encode('utf-8')
buff.close()
for instance in instance_qs:
output = heartbeat_stdout
task_result = {}
try:
with open(os.path.join(facts_path, instance.hostname), 'r') as facts_data:
output = facts_data.read()
task_result = json.loads(output)
except Exception:
logger.exception('Failed to read status from isolated instances, output:\n {}'.format(output))
if 'awx_capacity_cpu' in task_result and 'awx_capacity_mem' in task_result:
task_result = {
'capacity_cpu': task_result['awx_capacity_cpu'],
'capacity_mem': task_result['awx_capacity_mem'],
'version': task_result['awx_capacity_version']
}
cls.update_capacity(instance, task_result, awx_application_version)
logger.debug('Isolated instance {} successful heartbeat'.format(instance.hostname))
elif instance.capacity == 0:
logger.debug('Isolated instance {} previously marked as lost, could not re-join.'.format(
instance.hostname))
else:
logger.warning('Could not update status of isolated instance {}'.format(instance.hostname))
if instance.is_lost(isolated=True):
instance.capacity = 0
instance.save(update_fields=['capacity'])
logger.error('Isolated instance {} last checked in at {}, marked as lost.'.format(
instance.hostname, instance.modified))
finally:
if os.path.exists(facts_path):
shutil.rmtree(facts_path)
@staticmethod
def get_stdout_handle(instance, private_data_dir, event_data_key='job_id'):
dispatcher = CallbackQueueDispatcher()
def job_event_callback(event_data):
event_data.setdefault(event_data_key, instance.id)
if 'uuid' in event_data:
filename = '{}-partial.json'.format(event_data['uuid'])
partial_filename = os.path.join(private_data_dir, 'artifacts', 'job_events', filename)
try:
with codecs.open(partial_filename, 'r', encoding='utf-8') as f:
partial_event_data = json.load(f)
event_data.update(partial_event_data)
except IOError:
if event_data.get('event', '') != 'verbose':
logger.error('Missing callback data for event type `{}`, uuid {}, job {}.\nevent_data: {}'.format(
event_data.get('event', ''), event_data['uuid'], instance.id, event_data))
dispatcher.dispatch(event_data)
return OutputEventFilter(job_event_callback)
def run(self, instance, private_data_dir, proot_temp_dir):
"""
Run a job on an isolated host.
:param instance: a `model.Job` instance
:param private_data_dir: an absolute path on the local file system
where job-specific data should be written
(i.e., `/tmp/ansible_awx_xyz/`)
:param proot_temp_dir: a temporary directory which bwrap maps
restricted paths to
For a completed job run, this function returns (status, rc),
representing the status and return code of the isolated
`ansible-playbook` run.
"""
self.instance = instance
self.host = instance.execution_node
self.private_data_dir = private_data_dir
self.proot_temp_dir = proot_temp_dir
status, rc = self.dispatch()
if status == 'successful':
status, rc = self.check()
self.cleanup()
return status, rc

View File

@@ -1,333 +0,0 @@
#! /usr/bin/env python
import argparse
import base64
import codecs
import collections
import logging
import json
import os
import stat
import pipes
import re
import signal
import sys
import threading
import time
try:
from io import StringIO
except ImportError:
from StringIO import StringIO
import pexpect
import psutil
logger = logging.getLogger('awx.main.utils.expect')
def args2cmdline(*args):
return ' '.join([pipes.quote(a) for a in args])
def wrap_args_with_ssh_agent(args, ssh_key_path, ssh_auth_sock=None, silence_ssh_add=False):
if ssh_key_path:
ssh_add_command = args2cmdline('ssh-add', ssh_key_path)
if silence_ssh_add:
ssh_add_command = ' '.join([ssh_add_command, '2>/dev/null'])
cmd = ' && '.join([ssh_add_command,
args2cmdline('rm', '-f', ssh_key_path),
args2cmdline(*args)])
args = ['ssh-agent']
if ssh_auth_sock:
args.extend(['-a', ssh_auth_sock])
args.extend(['sh', '-c', cmd])
return args
def open_fifo_write(path, data):
'''open_fifo_write opens the fifo named pipe in a new thread.
This blocks the thread until an external process (such as ssh-agent)
reads data from the pipe.
'''
os.mkfifo(path, 0o600)
threading.Thread(
target=lambda p, d: open(p, 'w').write(d),
args=(path, data)
).start()
def run_pexpect(args, cwd, env, logfile,
cancelled_callback=None, expect_passwords={},
extra_update_fields=None, idle_timeout=None, job_timeout=0,
pexpect_timeout=5, proot_cmd='bwrap'):
'''
Run the given command using pexpect to capture output and provide
passwords when requested.
:param args: a list of `subprocess.call`-style arguments
representing a subprocess e.g., ['ls', '-la']
:param cwd: the directory in which the subprocess should
run
:param env: a dict containing environment variables for the
subprocess, ala `os.environ`
:param logfile: a file-like object for capturing stdout
:param cancelled_callback: a callable - which returns `True` or `False`
- signifying if the job has been prematurely
cancelled
:param expect_passwords: a dict of regular expression password prompts
to input values, i.e., {r'Password:*?$':
'some_password'}
:param extra_update_fields: a dict used to specify DB fields which should
be updated on the underlying model
object after execution completes
:param idle_timeout a timeout (in seconds); if new output is not
sent to stdout in this interval, the process
will be terminated
:param job_timeout a timeout (in seconds); if the total job runtime
exceeds this, the process will be killed
:param pexpect_timeout a timeout (in seconds) to wait on
`pexpect.spawn().expect()` calls
:param proot_cmd the command used to isolate processes, `bwrap`
Returns a tuple (status, return_code) i.e., `('successful', 0)`
'''
expect_passwords[pexpect.TIMEOUT] = None
expect_passwords[pexpect.EOF] = None
if not isinstance(expect_passwords, collections.OrderedDict):
# We iterate over `expect_passwords.keys()` and
# `expect_passwords.values()` separately to map matched inputs to
# patterns and choose the proper string to send to the subprocess;
# enforce usage of an OrderedDict so that the ordering of elements in
# `keys()` matches `values()`.
expect_passwords = collections.OrderedDict(expect_passwords)
password_patterns = list(expect_passwords.keys())
password_values = list(expect_passwords.values())
child = pexpect.spawn(
args[0], args[1:], cwd=cwd, env=env, ignore_sighup=True,
encoding='utf-8', echo=False, use_poll=True
)
child.logfile_read = logfile
canceled = False
timed_out = False
errored = False
last_stdout_update = time.time()
job_start = time.time()
while child.isalive():
result_id = child.expect(password_patterns, timeout=pexpect_timeout, searchwindowsize=100)
password = password_values[result_id]
if password is not None:
child.sendline(password)
last_stdout_update = time.time()
if cancelled_callback:
try:
canceled = cancelled_callback()
except Exception:
logger.exception('Could not check cancel callback - canceling immediately')
if isinstance(extra_update_fields, dict):
extra_update_fields['job_explanation'] = "System error during job execution, check system logs"
errored = True
else:
canceled = False
if not canceled and job_timeout != 0 and (time.time() - job_start) > job_timeout:
timed_out = True
if isinstance(extra_update_fields, dict):
extra_update_fields['job_explanation'] = "Job terminated due to timeout"
if canceled or timed_out or errored:
handle_termination(child.pid, child.args, proot_cmd, is_cancel=canceled)
if idle_timeout and (time.time() - last_stdout_update) > idle_timeout:
child.close(True)
canceled = True
if errored:
return 'error', child.exitstatus
elif canceled:
return 'canceled', child.exitstatus
elif child.exitstatus == 0 and not timed_out:
return 'successful', child.exitstatus
else:
return 'failed', child.exitstatus
def run_isolated_job(private_data_dir, secrets, logfile=sys.stdout):
'''
Launch `ansible-playbook`, executing a job packaged by
`build_isolated_job_data`.
:param private_data_dir: an absolute path on the local file system where
job metadata exists (i.e.,
`/tmp/ansible_awx_xyz/`)
:param secrets: a dict containing sensitive job metadata, {
'env': { ... } # environment variables,
'passwords': { ... } # pexpect password prompts
'ssh_key_data': 'RSA KEY DATA',
}
:param logfile: a file-like object for capturing stdout
Returns a tuple (status, return_code) i.e., `('successful', 0)`
'''
with open(os.path.join(private_data_dir, 'args'), 'r') as args:
args = json.load(args)
env = secrets.get('env', {})
expect_passwords = {
re.compile(pattern, re.M): password
for pattern, password in secrets.get('passwords', {}).items()
}
if 'AD_HOC_COMMAND_ID' in env:
cwd = private_data_dir
else:
cwd = os.path.join(private_data_dir, 'project')
# write the SSH key data into a fifo read by ssh-agent
ssh_key_data = secrets.get('ssh_key_data')
if ssh_key_data:
ssh_key_path = os.path.join(private_data_dir, 'ssh_key_data')
ssh_auth_sock = os.path.join(private_data_dir, 'ssh_auth.sock')
open_fifo_write(ssh_key_path, ssh_key_data)
args = wrap_args_with_ssh_agent(args, ssh_key_path, ssh_auth_sock)
idle_timeout = secrets.get('idle_timeout', 10)
job_timeout = secrets.get('job_timeout', 10)
pexpect_timeout = secrets.get('pexpect_timeout', 5)
# Use local callback directory
callback_dir = os.getenv('AWX_LIB_DIRECTORY')
if callback_dir is None:
raise RuntimeError('Location for callbacks must be specified '
'by environment variable AWX_LIB_DIRECTORY.')
env['ANSIBLE_CALLBACK_PLUGINS'] = os.path.join(callback_dir, 'isolated_callbacks')
if 'AD_HOC_COMMAND_ID' in env:
env['ANSIBLE_STDOUT_CALLBACK'] = 'minimal'
else:
env['ANSIBLE_STDOUT_CALLBACK'] = 'awx_display'
env['AWX_ISOLATED_DATA_DIR'] = private_data_dir
env['PYTHONPATH'] = env.get('PYTHONPATH', '') + callback_dir + ':'
venv_path = env.get('VIRTUAL_ENV')
if venv_path and not os.path.exists(venv_path):
raise RuntimeError(
'a valid Python virtualenv does not exist at {}'.format(venv_path)
)
return run_pexpect(args, cwd, env, logfile,
expect_passwords=expect_passwords,
idle_timeout=idle_timeout,
job_timeout=job_timeout,
pexpect_timeout=pexpect_timeout)
def handle_termination(pid, args, proot_cmd, is_cancel=True):
'''
Terminate a subprocess spawned by `pexpect`.
:param pid: the process id of the running the job.
:param args: the args for the job, i.e., ['ansible-playbook', 'abc.yml']
:param proot_cmd the command used to isolate processes i.e., `bwrap`
:param is_cancel: flag showing whether this termination is caused by
instance's cancel_flag.
'''
try:
if sys.version_info > (3, 0):
used_proot = proot_cmd.encode('utf-8') in args
else:
used_proot = proot_cmd in ' '.join(args)
if used_proot:
if not psutil:
os.kill(pid, signal.SIGKILL)
else:
try:
main_proc = psutil.Process(pid=pid)
child_procs = main_proc.children(recursive=True)
for child_proc in child_procs:
os.kill(child_proc.pid, signal.SIGKILL)
os.kill(main_proc.pid, signal.SIGKILL)
except (TypeError, psutil.Error):
os.kill(pid, signal.SIGKILL)
else:
os.kill(pid, signal.SIGTERM)
time.sleep(3)
except OSError:
keyword = 'cancel' if is_cancel else 'timeout'
logger.warn("Attempted to %s already finished job, ignoring" % keyword)
def __run__(private_data_dir):
buff = StringIO()
with codecs.open(os.path.join(private_data_dir, 'env'), 'r', encoding='utf-8') as f:
for line in f:
buff.write(line)
artifacts_dir = os.path.join(private_data_dir, 'artifacts')
# Standard out directed to pickup location without event filtering applied
stdout_filename = os.path.join(artifacts_dir, 'stdout')
os.mknod(stdout_filename, stat.S_IFREG | stat.S_IRUSR | stat.S_IWUSR)
stdout_handle = codecs.open(stdout_filename, 'w', encoding='utf-8')
status, rc = run_isolated_job(
private_data_dir,
json.loads(base64.b64decode(buff.getvalue())),
stdout_handle
)
for filename, data in [
('status', status),
('rc', rc),
]:
artifact_path = os.path.join(private_data_dir, 'artifacts', filename)
os.mknod(artifact_path, stat.S_IFREG | stat.S_IRUSR | stat.S_IWUSR)
with open(artifact_path, 'w') as f:
f.write(str(data))
if __name__ == '__main__':
import awx
__version__ = awx.__version__
parser = argparse.ArgumentParser(description='manage a daemonized, isolated ansible playbook')
parser.add_argument('--version', action='version', version=__version__ + '-isolated')
parser.add_argument('command', choices=['start', 'stop', 'is-alive'])
parser.add_argument('private_data_dir')
args = parser.parse_args()
private_data_dir = args.private_data_dir
pidfile = os.path.join(private_data_dir, 'pid')
if args.command == 'start':
# create a file to log stderr in case the daemonized process throws
# an exception before it gets to `pexpect.spawn`
stderr_path = os.path.join(private_data_dir, 'artifacts', 'daemon.log')
if not os.path.exists(stderr_path):
os.mknod(stderr_path, stat.S_IFREG | stat.S_IRUSR | stat.S_IWUSR)
stderr = open(stderr_path, 'w+')
import daemon
from daemon.pidfile import TimeoutPIDLockFile
context = daemon.DaemonContext(
pidfile=TimeoutPIDLockFile(pidfile),
stderr=stderr
)
with context:
__run__(private_data_dir)
sys.exit(0)
try:
with open(pidfile, 'r') as f:
pid = int(f.readline())
except IOError:
sys.exit(1)
if args.command == 'stop':
try:
with open(os.path.join(private_data_dir, 'args'), 'r') as args:
handle_termination(pid, json.load(args), 'bwrap')
except IOError:
handle_termination(pid, [], 'bwrap')
elif args.command == 'is-alive':
try:
os.kill(pid, signal.SIG_DFL)
sys.exit(0)
except OSError:
sys.exit(1)

View File

@@ -11,6 +11,7 @@ from jinja2 import Environment, StrictUndefined
from jinja2.exceptions import UndefinedError, TemplateSyntaxError
# Django
import django
from django.core import exceptions as django_exceptions
from django.db.models.signals import (
post_save,
@@ -18,14 +19,16 @@ from django.db.models.signals import (
)
from django.db.models.signals import m2m_changed
from django.db import models
from django.db.models.fields.related import add_lazy_relation
from django.db.models.fields.related import lazy_related_operation
from django.db.models.fields.related_descriptors import (
ReverseOneToOneDescriptor,
ForwardManyToOneDescriptor,
ManyToManyDescriptor,
ReverseManyToOneDescriptor,
create_forward_many_to_many_manager
)
from django.utils.encoding import smart_text
from django.utils.functional import cached_property
from django.utils.translation import ugettext_lazy as _
# jsonschema
@@ -43,14 +46,17 @@ from rest_framework import serializers
from awx.main.utils.filters import SmartFilter
from awx.main.utils.encryption import encrypt_value, decrypt_value, get_encryption_key
from awx.main.validators import validate_ssh_private_key
from awx.main.models.rbac import batch_role_ancestor_rebuilding, Role
from awx.main.models.rbac import (
batch_role_ancestor_rebuilding, Role,
ROLE_SINGLETON_SYSTEM_ADMINISTRATOR, ROLE_SINGLETON_SYSTEM_AUDITOR
)
from awx.main.constants import ENV_BLACKLIST
from awx.main import utils
__all__ = ['AutoOneToOneField', 'ImplicitRoleField', 'JSONField',
'SmartFilterField', 'update_role_parentage_for_instance',
'is_implicit_parent']
'SmartFilterField', 'OrderedManyToManyField',
'update_role_parentage_for_instance', 'is_implicit_parent']
# Provide a (better) custom error message for enum jsonschema validation
@@ -159,6 +165,13 @@ def is_implicit_parent(parent_role, child_role):
the model definition. This does not include any role parents that
might have been set by the user.
'''
if child_role.content_object is None:
# The only singleton implicit parent is the system admin being
# a parent of the system auditor role
return bool(
child_role.singleton_name == ROLE_SINGLETON_SYSTEM_AUDITOR and
parent_role.singleton_name == ROLE_SINGLETON_SYSTEM_ADMINISTRATOR
)
# Get the list of implicit parents that were defined at the class level.
implicit_parents = getattr(
child_role.content_object.__class__, child_role.role_field
@@ -217,6 +230,7 @@ class ImplicitRoleField(models.ForeignKey):
kwargs.setdefault('related_name', '+')
kwargs.setdefault('null', 'True')
kwargs.setdefault('editable', False)
kwargs.setdefault('on_delete', models.CASCADE)
super(ImplicitRoleField, self).__init__(*args, **kwargs)
def deconstruct(self):
@@ -234,7 +248,9 @@ class ImplicitRoleField(models.ForeignKey):
post_save.connect(self._post_save, cls, True, dispatch_uid='implicit-role-post-save')
post_delete.connect(self._post_delete, cls, True, dispatch_uid='implicit-role-post-delete')
add_lazy_relation(cls, self, "self", self.bind_m2m_changed)
function = lambda local, related, field: self.bind_m2m_changed(field, related, local)
lazy_related_operation(function, cls, "self", field=self)
def bind_m2m_changed(self, _self, _role_class, cls):
if not self.parent_role:
@@ -480,6 +496,86 @@ def format_ssh_private_key(value):
return True
@JSONSchemaField.format_checker.checks('url')
def format_url(value):
try:
parsed = urllib.parse.urlparse(value)
except Exception as e:
raise jsonschema.exceptions.FormatError(str(e))
if parsed.scheme == '':
raise jsonschema.exceptions.FormatError(
'Invalid URL: Missing url scheme (http, https, etc.)'
)
if parsed.netloc == '':
raise jsonschema.exceptions.FormatError(
'Invalid URL: {}'.format(value)
)
return True
class DynamicCredentialInputField(JSONSchemaField):
"""
Used to validate JSON for
`awx.main.models.credential:CredentialInputSource().metadata`.
Metadata for input sources is represented as a dictionary e.g.,
{'secret_path': '/kv/somebody', 'secret_key': 'password'}
For the data to be valid, the keys of this dictionary should correspond
with the metadata field (and datatypes) defined in the associated
target CredentialType e.g.,
"""
def schema(self, credential_type):
# determine the defined fields for the associated credential type
properties = {}
for field in credential_type.inputs.get('metadata', []):
field = field.copy()
properties[field['id']] = field
if field.get('choices', []):
field['enum'] = list(field['choices'])[:]
return {
'type': 'object',
'properties': properties,
'additionalProperties': False,
}
def validate(self, value, model_instance):
if not isinstance(value, dict):
return super(DynamicCredentialInputField, self).validate(value, model_instance)
super(JSONSchemaField, self).validate(value, model_instance)
credential_type = model_instance.source_credential.credential_type
errors = {}
for error in Draft4Validator(
self.schema(credential_type),
format_checker=self.format_checker
).iter_errors(value):
if error.validator == 'pattern' and 'error' in error.schema:
error.message = error.schema['error'].format(instance=error.instance)
if 'id' not in error.schema:
# If the error is not for a specific field, it's specific to
# `inputs` in general
raise django_exceptions.ValidationError(
error.message,
code='invalid',
params={'value': value},
)
errors[error.schema['id']] = [error.message]
defined_metadata = [field.get('id') for field in credential_type.inputs.get('metadata', [])]
for field in credential_type.inputs.get('required', []):
if field in defined_metadata and not value.get(field, None):
errors[field] = [_('required for %s') % (
credential_type.name
)]
if errors:
raise serializers.ValidationError({
'metadata': errors
})
class CredentialInputField(JSONSchemaField):
"""
Used to validate JSON for
@@ -542,7 +638,7 @@ class CredentialInputField(JSONSchemaField):
v != '$encrypted$',
model_instance.pk
]):
if not isinstance(getattr(model_instance, k), str):
if not isinstance(model_instance.inputs.get(k), str):
raise django_exceptions.ValidationError(
_('secret values must be of type string, not {}').format(type(v).__name__),
code='invalid',
@@ -592,18 +688,13 @@ class CredentialInputField(JSONSchemaField):
)
errors[error.schema['id']] = [error.message]
inputs = model_instance.credential_type.inputs
for field in inputs.get('required', []):
if not value.get(field, None):
errors[field] = [_('required for %s') % (
model_instance.credential_type.name
)]
defined_fields = model_instance.credential_type.defined_fields
# `ssh_key_unlock` requirements are very specific and can't be
# represented without complicated JSON schema
if (
model_instance.credential_type.managed_by_tower is True and
'ssh_key_unlock' in model_instance.credential_type.defined_fields
'ssh_key_unlock' in defined_fields
):
# in order to properly test the necessity of `ssh_key_unlock`, we
@@ -613,15 +704,15 @@ class CredentialInputField(JSONSchemaField):
# 'ssh_key_unlock': 'do-you-need-me?',
# }
# ...we have to fetch the actual key value from the database
if model_instance.pk and model_instance.ssh_key_data == '$encrypted$':
model_instance.ssh_key_data = model_instance.__class__.objects.get(
if model_instance.pk and model_instance.inputs.get('ssh_key_data') == '$encrypted$':
model_instance.inputs['ssh_key_data'] = model_instance.__class__.objects.get(
pk=model_instance.pk
).ssh_key_data
).inputs.get('ssh_key_data')
if model_instance.has_encrypted_ssh_key_data and not value.get('ssh_key_unlock'):
errors['ssh_key_unlock'] = [_('must be set when SSH key is encrypted.')]
if all([
model_instance.ssh_key_data,
model_instance.inputs.get('ssh_key_data'),
value.get('ssh_key_unlock'),
not model_instance.has_encrypted_ssh_key_data
]):
@@ -654,7 +745,7 @@ class CredentialTypeInputField(JSONSchemaField):
'type': 'object',
'properties': {
'type': {'enum': ['string', 'boolean']},
'format': {'enum': ['ssh_private_key']},
'format': {'enum': ['ssh_private_key', 'url']},
'choices': {
'type': 'array',
'minItems': 1,
@@ -671,6 +762,7 @@ class CredentialTypeInputField(JSONSchemaField):
'multiline': {'type': 'boolean'},
'secret': {'type': 'boolean'},
'ask_at_runtime': {'type': 'boolean'},
'default': {},
},
'additionalProperties': False,
'required': ['id', 'label'],
@@ -714,6 +806,14 @@ class CredentialTypeInputField(JSONSchemaField):
# If no type is specified, default to string
field['type'] = 'string'
if 'default' in field:
default = field['default']
_type = {'string': str, 'boolean': bool}[field['type']]
if type(default) != _type:
raise django_exceptions.ValidationError(
_('{} is not a {}').format(default, field['type'])
)
for key in ('choices', 'multiline', 'format', 'secret',):
if key in field and field['type'] != 'string':
raise django_exceptions.ValidationError(
@@ -890,3 +990,115 @@ class OAuth2ClientSecretField(models.CharField):
if value and value.startswith('$encrypted$'):
return decrypt_value(get_encryption_key('value', pk=None), value)
return value
class OrderedManyToManyDescriptor(ManyToManyDescriptor):
"""
Django doesn't seem to support:
class Meta:
ordering = [...]
...on custom through= relations for ManyToMany fields.
Meaning, queries made _through_ the intermediary table will _not_ apply an
ORDER_BY clause based on the `Meta.ordering` of the intermediary M2M class
(which is the behavior we want for "ordered" many to many relations):
https://github.com/django/django/blob/stable/1.11.x/django/db/models/fields/related_descriptors.py#L593
This descriptor automatically sorts all queries through this relation
using the `position` column on the M2M table.
"""
@cached_property
def related_manager_cls(self):
model = self.rel.related_model if self.reverse else self.rel.model
def add_custom_queryset_to_many_related_manager(many_related_manage_cls):
class OrderedManyRelatedManager(many_related_manage_cls):
def get_queryset(self):
return super(OrderedManyRelatedManager, self).get_queryset().order_by(
'%s__position' % self.through._meta.model_name
)
def add(self, *objs):
# Django < 2 doesn't support this method on
# ManyToManyFields w/ an intermediary model
# We should be able to remove this code snippet when we
# upgrade Django.
# see: https://github.com/django/django/blob/stable/1.11.x/django/db/models/fields/related_descriptors.py#L926
if not django.__version__.startswith('1.'):
raise RuntimeError(
'This method is no longer necessary in Django>=2'
)
try:
self.through._meta.auto_created = True
super(OrderedManyRelatedManager, self).add(*objs)
finally:
self.through._meta.auto_created = False
def remove(self, *objs):
# Django < 2 doesn't support this method on
# ManyToManyFields w/ an intermediary model
# We should be able to remove this code snippet when we
# upgrade Django.
# see: https://github.com/django/django/blob/stable/1.11.x/django/db/models/fields/related_descriptors.py#L944
if not django.__version__.startswith('1.'):
raise RuntimeError(
'This method is no longer necessary in Django>=2'
)
try:
self.through._meta.auto_created = True
super(OrderedManyRelatedManager, self).remove(*objs)
finally:
self.through._meta.auto_created = False
return OrderedManyRelatedManager
return add_custom_queryset_to_many_related_manager(
create_forward_many_to_many_manager(
model._default_manager.__class__,
self.rel,
reverse=self.reverse,
)
)
class OrderedManyToManyField(models.ManyToManyField):
"""
A ManyToManyField that automatically sorts all querysets
by a special `position` column on the M2M table
"""
def _update_m2m_position(self, sender, **kwargs):
if kwargs.get('action') in ('post_add', 'post_remove'):
order_with_respect_to = None
for field in sender._meta.local_fields:
if (
isinstance(field, models.ForeignKey) and
isinstance(kwargs['instance'], field.related_model)
):
order_with_respect_to = field.name
for i, ig in enumerate(sender.objects.filter(**{
order_with_respect_to: kwargs['instance'].pk}
)):
if ig.position != i:
ig.position = i
ig.save()
def contribute_to_class(self, cls, name, **kwargs):
super(OrderedManyToManyField, self).contribute_to_class(cls, name, **kwargs)
setattr(
cls, name,
OrderedManyToManyDescriptor(self.remote_field, reverse=False)
)
through = getattr(cls, name).through
if isinstance(through, str) and "." not in through:
# support lazy loading of string model names
through = '.'.join([cls._meta.app_label, through])
m2m_changed.connect(
self._update_m2m_position,
sender=through
)

View File

View File

@@ -0,0 +1,404 @@
import fnmatch
import json
import os
import shutil
import stat
import tempfile
import time
import logging
from django.conf import settings
import ansible_runner
import awx
from awx.main.utils import get_system_task_capacity
from awx.main.queue import CallbackQueueDispatcher
logger = logging.getLogger('awx.isolated.manager')
playbook_logger = logging.getLogger('awx.isolated.manager.playbooks')
def set_pythonpath(venv_libdir, env):
env.pop('PYTHONPATH', None) # default to none if no python_ver matches
for version in os.listdir(venv_libdir):
if fnmatch.fnmatch(version, 'python[23].*'):
if os.path.isdir(os.path.join(venv_libdir, version)):
env['PYTHONPATH'] = os.path.join(venv_libdir, version, "site-packages") + ":"
break
class IsolatedManager(object):
def __init__(self, cancelled_callback=None, check_callback=None):
"""
:param cancelled_callback: a callable - which returns `True` or `False`
- signifying if the job has been prematurely
cancelled
"""
self.cancelled_callback = cancelled_callback
self.check_callback = check_callback
self.idle_timeout = max(60, 2 * settings.AWX_ISOLATED_CONNECTION_TIMEOUT)
self.started_at = None
self.captured_command_artifact = False
def build_runner_params(self, hosts, verbosity=1):
env = dict(os.environ.items())
env['ANSIBLE_RETRY_FILES_ENABLED'] = 'False'
env['ANSIBLE_HOST_KEY_CHECKING'] = 'False'
env['ANSIBLE_LIBRARY'] = os.path.join(os.path.dirname(awx.__file__), 'plugins', 'isolated')
set_pythonpath(os.path.join(settings.ANSIBLE_VENV_PATH, 'lib'), env)
def finished_callback(runner_obj):
if runner_obj.status == 'failed' and runner_obj.config.playbook != 'check_isolated.yml':
# failed for clean_isolated.yml just means the playbook hasn't
# exited on the isolated host
stdout = runner_obj.stdout.read()
playbook_logger.error(stdout)
elif runner_obj.status == 'timeout':
# this means that the default idle timeout of
# (2 * AWX_ISOLATED_CONNECTION_TIMEOUT) was exceeded
# (meaning, we tried to sync with an isolated node, and we got
# no new output for 2 * AWX_ISOLATED_CONNECTION_TIMEOUT seconds)
# this _usually_ means SSH key auth from the controller ->
# isolated didn't work, and ssh is hung waiting on interactive
# input e.g.,
#
# awx@isolated's password:
stdout = runner_obj.stdout.read()
playbook_logger.error(stdout)
else:
playbook_logger.info(runner_obj.stdout.read())
inventory = '\n'.join([
'{} ansible_ssh_user={}'.format(host, settings.AWX_ISOLATED_USERNAME)
for host in hosts
])
return {
'project_dir': os.path.abspath(os.path.join(
os.path.dirname(awx.__file__),
'playbooks'
)),
'inventory': inventory,
'envvars': env,
'finished_callback': finished_callback,
'verbosity': verbosity,
'cancel_callback': self.cancelled_callback,
'settings': {
'idle_timeout': self.idle_timeout,
'job_timeout': settings.AWX_ISOLATED_LAUNCH_TIMEOUT,
'pexpect_timeout': getattr(settings, 'PEXPECT_TIMEOUT', 5),
'suppress_ansible_output': True,
},
}
def path_to(self, *args):
return os.path.join(self.private_data_dir, *args)
def run_management_playbook(self, playbook, private_data_dir, **kw):
iso_dir = tempfile.mkdtemp(
prefix=playbook,
dir=private_data_dir
)
params = self.runner_params.copy()
params['playbook'] = playbook
params['private_data_dir'] = iso_dir
params.update(**kw)
if all([
getattr(settings, 'AWX_ISOLATED_KEY_GENERATION', False) is True,
getattr(settings, 'AWX_ISOLATED_PRIVATE_KEY', None)
]):
params['ssh_key'] = settings.AWX_ISOLATED_PRIVATE_KEY
return ansible_runner.interface.run(**params)
def dispatch(self, playbook=None, module=None, module_args=None):
'''
Ship the runner payload to a remote host for isolated execution.
'''
self.handled_events = set()
self.started_at = time.time()
# exclude certain files from the rsync
rsync_exclude = [
# don't rsync source control metadata (it can be huge!)
'- /project/.git',
'- /project/.svn',
'- /project/.hg',
# don't rsync job events that are in the process of being written
'- /artifacts/job_events/*-partial.json.tmp',
# don't rsync the ssh_key FIFO
'- /env/ssh_key',
]
for filename, data in (
['.rsync-filter', '\n'.join(rsync_exclude)],
):
path = self.path_to(filename)
with open(path, 'w') as f:
f.write(data)
os.chmod(path, stat.S_IRUSR)
extravars = {
'src': self.private_data_dir,
'dest': settings.AWX_PROOT_BASE_PATH,
'ident': self.ident
}
if playbook:
extravars['playbook'] = playbook
if module and module_args:
extravars['module'] = module
extravars['module_args'] = module_args
logger.debug('Starting job {} on isolated host with `run_isolated.yml` playbook.'.format(self.instance.id))
runner_obj = self.run_management_playbook('run_isolated.yml',
self.private_data_dir,
extravars=extravars)
return runner_obj.status, runner_obj.rc
def check(self, interval=None):
"""
Repeatedly poll the isolated node to determine if the job has run.
On success, copy job artifacts to the controlling node.
On failure, continue to poll the isolated node (until the job timeout
is exceeded).
For a completed job run, this function returns (status, rc),
representing the status and return code of the isolated
`ansible-playbook` run.
:param interval: an interval (in seconds) to wait between status polls
"""
interval = interval if interval is not None else settings.AWX_ISOLATED_CHECK_INTERVAL
extravars = {'src': self.private_data_dir}
status = 'failed'
rc = None
last_check = time.time()
dispatcher = CallbackQueueDispatcher()
while status == 'failed':
canceled = self.cancelled_callback() if self.cancelled_callback else False
if not canceled and time.time() - last_check < interval:
# If the job isn't cancelled, but we haven't waited `interval` seconds, wait longer
time.sleep(1)
continue
if canceled:
logger.warning('Isolated job {} was manually cancelled.'.format(self.instance.id))
logger.debug('Checking on isolated job {} with `check_isolated.yml`.'.format(self.instance.id))
runner_obj = self.run_management_playbook('check_isolated.yml',
self.private_data_dir,
extravars=extravars)
status, rc = runner_obj.status, runner_obj.rc
if self.check_callback is not None and not self.captured_command_artifact:
command_path = self.path_to('artifacts', self.ident, 'command')
# If the configuration artifact has been synced back, update the model
if os.path.exists(command_path):
try:
with open(command_path, 'r') as f:
data = json.load(f)
self.check_callback(data)
self.captured_command_artifact = True
except json.decoder.JSONDecodeError: # Just in case it's not fully here yet.
pass
self.consume_events(dispatcher)
last_check = time.time()
if status == 'successful':
status_path = self.path_to('artifacts', self.ident, 'status')
rc_path = self.path_to('artifacts', self.ident, 'rc')
if os.path.exists(status_path):
with open(status_path, 'r') as f:
status = f.readline()
with open(rc_path, 'r') as f:
rc = int(f.readline())
else:
# if there's no status file, it means that runner _probably_
# exited with a traceback (which should be logged to
# daemon.log) Record it so we can see how runner failed.
daemon_path = self.path_to('daemon.log')
if os.path.exists(daemon_path):
with open(daemon_path, 'r') as f:
self.instance.result_traceback = f.read()
self.instance.save(update_fields=['result_traceback'])
else:
logger.error('Failed to rsync daemon.log (is ansible-runner installed on the isolated host?)')
status = 'failed'
rc = 1
# consume events one last time just to be sure we didn't miss anything
# in the final sync
self.consume_events(dispatcher)
# emit an EOF event
event_data = {
'event': 'EOF',
'final_counter': len(self.handled_events)
}
event_data.setdefault(self.event_data_key, self.instance.id)
dispatcher.dispatch(event_data)
return status, rc
def consume_events(self, dispatcher):
# discover new events and ingest them
events_path = self.path_to('artifacts', self.ident, 'job_events')
# it's possible that `events_path` doesn't exist *yet*, because runner
# hasn't actually written any events yet (if you ran e.g., a sleep 30)
# only attempt to consume events if any were rsynced back
if os.path.exists(events_path):
for event in set(os.listdir(events_path)) - self.handled_events:
path = os.path.join(events_path, event)
if os.path.exists(path):
try:
event_data = json.load(
open(os.path.join(events_path, event), 'r')
)
except json.decoder.JSONDecodeError:
# This means the event we got back isn't valid JSON
# that can happen if runner is still partially
# writing an event file while it's rsyncing
# these event writes are _supposed_ to be atomic
# but it doesn't look like they actually are in
# practice
# in this scenario, just ignore this event and try it
# again on the next sync
pass
event_data.setdefault(self.event_data_key, self.instance.id)
dispatcher.dispatch(event_data)
self.handled_events.add(event)
# handle artifacts
if event_data.get('event_data', {}).get('artifact_data', {}):
self.instance.artifacts = event_data['event_data']['artifact_data']
self.instance.save(update_fields=['artifacts'])
def cleanup(self):
# If the job failed for any reason, make a last-ditch effort at cleanup
extravars = {
'private_data_dir': self.private_data_dir,
'cleanup_dirs': [
self.private_data_dir,
],
}
logger.debug('Cleaning up job {} on isolated host with `clean_isolated.yml` playbook.'.format(self.instance.id))
self.run_management_playbook(
'clean_isolated.yml',
self.private_data_dir,
extravars=extravars
)
@classmethod
def update_capacity(cls, instance, task_result):
instance.version = 'ansible-runner-{}'.format(task_result['version'])
if instance.capacity == 0 and task_result['capacity_cpu']:
logger.warning('Isolated instance {} has re-joined.'.format(instance.hostname))
instance.cpu = int(task_result['cpu'])
instance.memory = int(task_result['mem'])
instance.cpu_capacity = int(task_result['capacity_cpu'])
instance.mem_capacity = int(task_result['capacity_mem'])
instance.capacity = get_system_task_capacity(scale=instance.capacity_adjustment,
cpu_capacity=int(task_result['capacity_cpu']),
mem_capacity=int(task_result['capacity_mem']))
instance.save(update_fields=['cpu', 'memory', 'cpu_capacity', 'mem_capacity', 'capacity', 'version', 'modified'])
def health_check(self, instance_qs):
'''
:param instance_qs: List of Django objects representing the
isolated instances to manage
Runs playbook that will
- determine if instance is reachable
- find the instance capacity
- clean up orphaned private files
Performs save on each instance to update its capacity.
'''
instance_qs = [i for i in instance_qs if i.enabled]
if not len(instance_qs):
return
try:
private_data_dir = tempfile.mkdtemp(
prefix='awx_iso_heartbeat_',
dir=settings.AWX_PROOT_BASE_PATH
)
self.runner_params = self.build_runner_params([
instance.hostname for instance in instance_qs
])
self.runner_params['private_data_dir'] = private_data_dir
self.runner_params['forks'] = len(instance_qs)
runner_obj = self.run_management_playbook(
'heartbeat_isolated.yml',
private_data_dir
)
if runner_obj.status == 'successful':
for instance in instance_qs:
task_result = {}
try:
task_result = runner_obj.get_fact_cache(instance.hostname)
except Exception:
logger.exception('Failed to read status from isolated instances')
if 'awx_capacity_cpu' in task_result and 'awx_capacity_mem' in task_result:
task_result = {
'cpu': task_result['awx_cpu'],
'mem': task_result['awx_mem'],
'capacity_cpu': task_result['awx_capacity_cpu'],
'capacity_mem': task_result['awx_capacity_mem'],
'version': task_result['awx_capacity_version']
}
IsolatedManager.update_capacity(instance, task_result)
logger.debug('Isolated instance {} successful heartbeat'.format(instance.hostname))
elif instance.capacity == 0:
logger.debug('Isolated instance {} previously marked as lost, could not re-join.'.format(
instance.hostname))
else:
logger.warning('Could not update status of isolated instance {}'.format(instance.hostname))
if instance.is_lost(isolated=True):
instance.capacity = 0
instance.save(update_fields=['capacity'])
logger.error('Isolated instance {} last checked in at {}, marked as lost.'.format(
instance.hostname, instance.modified))
finally:
if os.path.exists(private_data_dir):
shutil.rmtree(private_data_dir)
def run(self, instance, private_data_dir, playbook, module, module_args,
event_data_key, ident=None):
"""
Run a job on an isolated host.
:param instance: a `model.Job` instance
:param private_data_dir: an absolute path on the local file system
where job-specific data should be written
(i.e., `/tmp/awx_N_xyz/`)
:param playbook: the playbook to run
:param module: the module to run
:param module_args: the module args to use
:param event_data_key: e.g., job_id, inventory_id, ...
For a completed job run, this function returns (status, rc),
representing the status and return code of the isolated
`ansible-playbook` run.
"""
self.ident = ident
self.event_data_key = event_data_key
self.instance = instance
self.private_data_dir = private_data_dir
self.runner_params = self.build_runner_params(
[instance.execution_node],
verbosity=min(5, self.instance.verbosity)
)
status, rc = self.dispatch(playbook, module, module_args)
if status == 'successful':
status, rc = self.check()
else:
# emit an EOF event
event_data = {'event': 'EOF', 'final_counter': 0}
event_data.setdefault(self.event_data_key, self.instance.id)
CallbackQueueDispatcher().dispatch(event_data)
return status, rc

View File

@@ -59,7 +59,7 @@ class Command(BaseCommand):
if len(pks_to_delete):
ActivityStream.objects.filter(pk__in=pks_to_delete).delete()
n_deleted_items += len(pks_to_delete)
self.logger.log(99, "Removed %d items", n_deleted_items)
self.logger.info("Removed {} items".format(n_deleted_items))
def handle(self, *args, **options):
self.verbosity = int(options.get('verbosity', 1))

View File

@@ -1,148 +0,0 @@
# Copyright (c) 2015 Ansible, Inc.
# All Rights Reserved
# Python
import re
import sys
from dateutil.relativedelta import relativedelta
# Django
from django.core.management.base import BaseCommand, CommandError
from django.db import transaction
from django.utils.timezone import now
# AWX
from awx.main.models.fact import Fact
from awx.conf.license import feature_enabled
OLDER_THAN = 'older_than'
GRANULARITY = 'granularity'
class CleanupFacts(object):
def __init__(self):
self.timestamp = None
# Find all with timestamp < older_than
# Start search at < older_than, stop search at oldest entry
# Find all factVersion < pivot && > (pivot - granularity) grouped by host sorted by time descending (because it's indexed this way)
# foreach group
# Delete all except LAST entry (or Delete all except the FIRST entry, it's an arbitrary decision)
#
# pivot -= granularity
# group by host
def cleanup(self, older_than_abs, granularity, module=None):
fact_oldest = Fact.objects.all().order_by('timestamp').first()
if not fact_oldest:
return 0
kv = {
'timestamp__lte': older_than_abs
}
if module:
kv['module'] = module
# Special case, granularity=0x where x is d, w, or y
# The intent is to delete all facts < older_than_abs
if granularity == relativedelta():
qs = Fact.objects.filter(**kv)
count = qs.count()
qs.delete()
return count
total = 0
date_pivot = older_than_abs
while date_pivot > fact_oldest.timestamp:
date_pivot_next = date_pivot - granularity
# For the current time window.
# Delete all facts expect the fact that matches the largest timestamp.
kv = {
'timestamp__lte': date_pivot
}
if module:
kv['module'] = module
fact_version_obj = Fact.objects.filter(**kv).order_by('-timestamp').first()
if fact_version_obj:
kv = {
'timestamp__lt': fact_version_obj.timestamp,
'timestamp__gt': date_pivot_next
}
if module:
kv['module'] = module
qs = Fact.objects.filter(**kv)
count = qs.count()
qs.delete()
total += count
date_pivot = date_pivot_next
return total
'''
older_than and granularity are of type relativedelta
'''
def run(self, older_than, granularity, module=None):
t = now()
deleted_count = self.cleanup(t - older_than, granularity, module=module)
print("Deleted %d facts." % deleted_count)
class Command(BaseCommand):
help = 'Cleanup facts. For each host older than the value specified, keep one fact scan for each time window (granularity).'
def add_arguments(self, parser):
parser.add_argument('--older_than',
dest='older_than',
default='30d',
help='Specify the relative time to consider facts older than (w)eek (d)ay or (y)ear (i.e. 5d, 2w, 1y). Defaults to 30d.')
parser.add_argument('--granularity',
dest='granularity',
default='1w',
help='Window duration to group same hosts by for deletion (w)eek (d)ay or (y)ear (i.e. 5d, 2w, 1y). Defaults to 1w.')
parser.add_argument('--module',
dest='module',
default=None,
help='Limit cleanup to a particular module.')
def __init__(self):
super(Command, self).__init__()
def string_time_to_timestamp(self, time_string):
units = {
'y': 'years',
'd': 'days',
'w': 'weeks',
'm': 'months'
}
try:
match = re.match(r'(?P<value>[0-9]+)(?P<unit>.*)', time_string)
group = match.groupdict()
kv = {}
units_verbose = units[group['unit']]
kv[units_verbose]= int(group['value'])
return relativedelta(**kv)
except (KeyError, TypeError, AttributeError):
return None
@transaction.atomic
def handle(self, *args, **options):
sys.stderr.write("This command has been deprecated and will be removed in a future release.\n")
if not feature_enabled('system_tracking'):
raise CommandError("The System Tracking feature is not enabled for your instance")
cleanup_facts = CleanupFacts()
if not all([options[GRANULARITY], options[OLDER_THAN]]):
raise CommandError('Both --granularity and --older_than are required.')
older_than = self.string_time_to_timestamp(options[OLDER_THAN])
granularity = self.string_time_to_timestamp(options[GRANULARITY])
if older_than is None:
raise CommandError('--older_than invalid value "%s"' % options[OLDER_THAN])
if granularity is None:
raise CommandError('--granularity invalid value "%s"' % options[GRANULARITY])
cleanup_facts.run(older_than, granularity, module=options['module'])

View File

@@ -0,0 +1,25 @@
import logging
from django.core import management
from django.core.management.base import BaseCommand
from django.contrib.sessions.models import Session
class Command(BaseCommand):
def init_logging(self):
log_levels = dict(enumerate([logging.ERROR, logging.INFO,
logging.DEBUG, 0]))
self.logger = logging.getLogger('awx.main.commands.cleanup_sessions')
self.logger.setLevel(log_levels.get(self.verbosity, 0))
handler = logging.StreamHandler()
handler.setFormatter(logging.Formatter('%(message)s'))
self.logger.addHandler(handler)
self.logger.propagate = False
def execute(self, *args, **options):
self.verbosity = int(options.get('verbosity', 1))
self.init_logging()
total_sessions = Session.objects.all().count()
management.call_command('clearsessions')
self.logger.info("Expired Sessions deleted {}".format(total_sessions - Session.objects.all().count()))

View File

@@ -0,0 +1,28 @@
import logging
from django.core import management
from django.core.management.base import BaseCommand
from awx.main.models import OAuth2AccessToken
from oauth2_provider.models import RefreshToken
class Command(BaseCommand):
def init_logging(self):
log_levels = dict(enumerate([logging.ERROR, logging.INFO,
logging.DEBUG, 0]))
self.logger = logging.getLogger('awx.main.commands.cleanup_tokens')
self.logger.setLevel(log_levels.get(self.verbosity, 0))
handler = logging.StreamHandler()
handler.setFormatter(logging.Formatter('%(message)s'))
self.logger.addHandler(handler)
self.logger.propagate = False
def execute(self, *args, **options):
self.verbosity = int(options.get('verbosity', 1))
self.init_logging()
total_accesstokens = OAuth2AccessToken.objects.all().count()
total_refreshtokens = RefreshToken.objects.all().count()
management.call_command('cleartokens')
self.logger.info("Expired OAuth 2 Access Tokens deleted: {}".format(total_accesstokens - OAuth2AccessToken.objects.all().count()))
self.logger.info("Expired OAuth 2 Refresh Tokens deleted: {}".format(total_refreshtokens - RefreshToken.objects.all().count()))

View File

@@ -34,7 +34,7 @@ class Command(BaseCommand):
scm_update_cache_timeout=0,
organization=o)
p.save(skip_update=True)
ssh_type = CredentialType.from_v1_kind('ssh')
ssh_type = CredentialType.objects.filter(namespace='ssh').first()
c = Credential.objects.create(credential_type=ssh_type,
name='Demo Credential',
inputs={
@@ -47,7 +47,7 @@ class Command(BaseCommand):
created_by=superuser)
Host.objects.create(name='localhost',
inventory=i,
variables="ansible_connection: local",
variables="ansible_connection: local\nansible_python_interpreter: '{{ ansible_playbook_python }}'",
created_by=superuser)
jt = JobTemplate.objects.create(name='Demo Job Template',
playbook='hello_world.yml',

View File

@@ -4,8 +4,6 @@ from importlib import import_module
# Django
from django.utils import timezone
from django.conf import settings
from django.contrib.auth import logout
from django.http import HttpRequest
from django.core.management.base import BaseCommand, CommandError
from django.contrib.auth.models import User
from django.contrib.sessions.models import Session
@@ -29,9 +27,9 @@ class Command(BaseCommand):
# with consideration for timezones.
start = timezone.now()
sessions = Session.objects.filter(expire_date__gte=start).iterator()
request = HttpRequest()
for session in sessions:
user_id = session.get_decoded().get('_auth_user_id')
if (user is None) or (user_id and user.id == int(user_id)):
request.session = import_module(settings.SESSION_ENGINE).SessionStore(session.session_key)
logout(request)
session = import_module(settings.SESSION_ENGINE).SessionStore(session.session_key)
# Log out the session, but without the need for a request object.
session.flush()

View File

@@ -0,0 +1,31 @@
import logging
from awx.main.analytics import gather, ship
from django.core.management.base import BaseCommand
class Command(BaseCommand):
'''
Gather AWX analytics data
'''
help = 'Gather AWX analytics data'
def add_arguments(self, parser):
parser.add_argument('--ship', dest='ship', action='store_true',
help='Enable to ship metrics via insights-client')
def init_logging(self):
self.logger = logging.getLogger('awx.main.analytics')
handler = logging.StreamHandler()
handler.setLevel(logging.DEBUG)
handler.setFormatter(logging.Formatter('%(message)s'))
self.logger.addHandler(handler)
self.logger.propagate = False
def handle(self, *args, **options):
tgz = gather()
self.init_logging()
if tgz:
self.logger.debug(tgz)
if options.get('ship'):
ship(tgz)

Some files were not shown because too many files have changed in this diff Show More