make obvious updates to the logging related docs

This commit is contained in:
AlanCoding
2017-02-09 12:37:57 -05:00
parent 1c6f037076
commit 279546e110
2 changed files with 52 additions and 17 deletions

View File

@@ -26,7 +26,7 @@ from the API. These data loggers are the following.
These loggers only use log-level of INFO. These loggers only use log-level of INFO.
Additionally, the standard Tower logs should be deliverable through this Additionally, the standard Tower logs are be deliverable through this
same mechanism. It should be obvious to the user how to enable to disable same mechanism. It should be obvious to the user how to enable to disable
each of these 5 sources of data without manipulating a complex dictionary each of these 5 sources of data without manipulating a complex dictionary
in their local settings file, as well as adjust the log-level consumed in their local settings file, as well as adjust the log-level consumed
@@ -34,16 +34,19 @@ from the standard Tower logs.
## Supported Services ## Supported Services
Currently committed to support: Committed to support:
- Splunk - Splunk
- Elastic Stack / ELK Stack / Elastic Cloud - Elastic Stack / ELK Stack / Elastic Cloud
Under consideration for testing: Have tested:
- Sumo Logic - Sumologic
- Datadog
- Loggly - Loggly
Considered, but have not tested:
- Datadog
- Red Hat Common Logging via logstash connector - Red Hat Common Logging via logstash connector
### Elastic Search Instructions ### Elastic Search Instructions
@@ -64,8 +67,8 @@ make docker-compose-elk
make docker-compose-cluster-elk make docker-compose-cluster-elk
``` ```
Kibana is the visualization service, and it can be accessed in a web browser For more instructions on getting started with the environment this stands
by going to `{server address}:5601`. up, also refer to instructions in `/tools/elastic/README.md`.
If you were to start from scratch, standing up your own version the elastic If you were to start from scratch, standing up your own version the elastic
stack, then the only change you should need is to add the following lines stack, then the only change you should need is to add the following lines
@@ -149,6 +152,8 @@ the job model.
In addition to the common fields, this will contain a `msg` field with In addition to the common fields, this will contain a `msg` field with
the log message. Errors contain a separate `traceback` field. the log message. Errors contain a separate `traceback` field.
These logs can be enabled or disabled in CTiT by adding or removing
it to the setting `LOG_AGGREGATOR_LOGGERS`.
# Configuring Inside of Tower # Configuring Inside of Tower
@@ -158,10 +163,12 @@ supported services:
- Host - Host
- Port - Port
- some kind of token - The type of service, allowing service-specific customizations
- enabling sending logs, and selecting which loggers to send - Optional username for the connection, used by certain services
- use fully qualified domain name (fqdn) or not - Some kind of token or password
- flag to use HTTPS or not - A flag to indicate how system tracking records will be sent
- Selecting which loggers to send
- Enabling sending logs
Some settings for the log handler will not be exposed to the user via Some settings for the log handler will not be exposed to the user via
this mechanism. In particular, threading (enabled), and connection type this mechanism. In particular, threading (enabled), and connection type

View File

@@ -1,8 +1,7 @@
# Docker ELK / Elastic Stack Development Tools # Docker ELK / Elastic Stack Development Tools
These are tools to run a containerized version of ELK stack, comprising These are tools to run a containerized version of ELK stack, comprising
of Logstash, Elastic Search, and Kibana. There are also cases where of Logstash, Elastic Search, and Kibana.
only a subset of these are needed to run.
A copy of the license is in `docs/licenses/docker-elk.txt` A copy of the license is in `docs/licenses/docker-elk.txt`
@@ -12,12 +11,10 @@ Due to complex requirements from the elastic search container upstream, there
is a prerequisite to get the containers running. The docker _host_ machine is a prerequisite to get the containers running. The docker _host_ machine
must have the `max_map_count` variable increased. For a developer using must have the `max_map_count` variable increased. For a developer using
docker-machine with something like VirtualBox of VMWare, this can be docker-machine with something like VirtualBox of VMWare, this can be
done by getting bash in the running Docker machine. Example: done by getting via bash in the running Docker machine. Example:
```bash ```bash
$ docker-machine ssh default docker-machine ssh default sudo sysctl -w vm.max_map_count=262144
docker@default:~$ sudo sysctl -w vm.max_map_count=262144
vm.max_map_count = 262144
``` ```
After this, the containers can be started up with commands like: After this, the containers can be started up with commands like:
@@ -32,6 +29,37 @@ make docker-compose-cluster-elk
These are ran from the root folder of the ansible-tower repository. These are ran from the root folder of the ansible-tower repository.
Kibana is the visualization service, and it can be accessed in a web browser
by going to `{server address}:5601`.
### Authentication
The default logstash configuration makes use of basic auth, so a username
and password is needed in the configuration, in addition to the other
parameters. The following settings are supported:
```
{
"LOG_AGGREGATOR_HOST": "logstash",
"LOG_AGGREGATOR_PORT": 8085,
"LOG_AGGREGATOR_TYPE": "logstash",
"LOG_AGGREGATOR_USERNAME": "awx_logger",
"LOG_AGGREGATOR_PASSWORD": "workflows",
"LOG_AGGREGATOR_LOGGERS": [
"awx",
"activity_stream",
"job_events",
"system_tracking"
],
"LOG_AGGREGATOR_INDIVIDUAL_FACTS": false,
"LOG_AGGREGATOR_ENABLED": true
}
```
These can be entered via Configure-Tower-in-Tower by making a POST to
`/api/v1/settings/logging/`.
### Connecting Logstash to 3rd Party Receivers ### Connecting Logstash to 3rd Party Receivers
In order to send these logs to an external consumer of logstash format In order to send these logs to an external consumer of logstash format