mirror of
https://github.com/ansible/awx.git
synced 2026-02-25 06:56:00 -03:30
remove unused log handlers and cleanup
This commit is contained in:
@@ -1,60 +1,24 @@
|
|||||||
# Copyright (c) 2017 Ansible Tower by Red Hat
|
# Copyright (c) 2017 Ansible Tower by Red Hat
|
||||||
# All Rights Reserved.
|
# All Rights Reserved.
|
||||||
|
|
||||||
# common
|
# Python
|
||||||
import socket
|
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
# Splunk
|
|
||||||
import urllib
|
|
||||||
import json
|
import json
|
||||||
import gzip
|
|
||||||
import cStringIO
|
|
||||||
|
|
||||||
import requests
|
import requests
|
||||||
|
from copy import copy
|
||||||
from .utils import parse_config_file, get_config_from_env
|
|
||||||
|
|
||||||
# loggly
|
# loggly
|
||||||
import traceback
|
import traceback
|
||||||
|
|
||||||
from requests_futures.sessions import FuturesSession
|
from requests_futures.sessions import FuturesSession
|
||||||
|
|
||||||
# Logstash
|
|
||||||
from logstash import formatter
|
|
||||||
|
|
||||||
# custom
|
# custom
|
||||||
from copy import copy
|
|
||||||
from requests.auth import HTTPBasicAuth
|
|
||||||
from django.conf import settings as django_settings
|
from django.conf import settings as django_settings
|
||||||
|
|
||||||
|
# AWX external logging handler, generally designed to be used
|
||||||
# Logstash
|
# with the accompanying LogstashHandler, derives from python-logstash library
|
||||||
# https://github.com/vklochan/python-logstash
|
# Non-blocking request accomplished by FuturesSession, similar
|
||||||
class TCPLogstashHandler(logging.handlers.SocketHandler, object):
|
# to the loggly-python-handler library (not used)
|
||||||
"""Python logging handler for Logstash. Sends events over TCP.
|
|
||||||
:param host: The host of the logstash server.
|
|
||||||
:param port: The port of the logstash server (default 5959).
|
|
||||||
:param message_type: The type of the message (default logstash).
|
|
||||||
:param fqdn; Indicates whether to show fully qualified domain name or not (default False).
|
|
||||||
:param version: version of logstash event schema (default is 0).
|
|
||||||
:param tags: list of tags for a logger (default is None).
|
|
||||||
"""
|
|
||||||
|
|
||||||
def __init__(self, host, port=5959, message_type='logstash', tags=None, fqdn=False, version=0):
|
|
||||||
super(TCPLogstashHandler, self).__init__(host, port)
|
|
||||||
if version == 1:
|
|
||||||
self.formatter = formatter.LogstashFormatterVersion1(message_type, tags, fqdn)
|
|
||||||
else:
|
|
||||||
self.formatter = formatter.LogstashFormatterVersion0(message_type, tags, fqdn)
|
|
||||||
|
|
||||||
def makePickle(self, record):
|
|
||||||
return self.formatter.format(record) + b'\n'
|
|
||||||
|
|
||||||
|
|
||||||
# techniquest borrowed from the loggly library
|
|
||||||
# https://github.com/varshneyjayant/loggly-python-handler
|
|
||||||
# MIT License
|
|
||||||
|
|
||||||
# Translation of parameter names to names in Django settings
|
# Translation of parameter names to names in Django settings
|
||||||
PARAM_NAMES = {
|
PARAM_NAMES = {
|
||||||
@@ -66,7 +30,6 @@ PARAM_NAMES = {
|
|||||||
'enabled_loggers': 'LOG_AGGREGATOR_LOGGERS',
|
'enabled_loggers': 'LOG_AGGREGATOR_LOGGERS',
|
||||||
'indv_facts': 'LOG_AGGREGATOR_INDIVIDUAL_FACTS',
|
'indv_facts': 'LOG_AGGREGATOR_INDIVIDUAL_FACTS',
|
||||||
}
|
}
|
||||||
# TODO: figure out what to do with LOG_AGGREGATOR_LOGGERS (if anything)
|
|
||||||
|
|
||||||
|
|
||||||
def unused_callback(sess, resp):
|
def unused_callback(sess, resp):
|
||||||
@@ -101,7 +64,7 @@ class HTTPSHandler(logging.Handler):
|
|||||||
if not self.username:
|
if not self.username:
|
||||||
# Logstash authentication not enabled
|
# Logstash authentication not enabled
|
||||||
return
|
return
|
||||||
logstash_auth = HTTPBasicAuth(self.username, self.password)
|
logstash_auth = requests.auth.HTTPBasicAuth(self.username, self.password)
|
||||||
self.session.auth = logstash_auth
|
self.session.auth = logstash_auth
|
||||||
elif self.message_type == 'splunk':
|
elif self.message_type == 'splunk':
|
||||||
auth_header = "Splunk %s" % self.password
|
auth_header = "Splunk %s" % self.password
|
||||||
@@ -111,17 +74,33 @@ class HTTPSHandler(logging.Handler):
|
|||||||
}
|
}
|
||||||
self.session.headers.update(headers)
|
self.session.headers.update(headers)
|
||||||
|
|
||||||
|
def get_http_host(self):
|
||||||
|
host = self.host
|
||||||
|
if not host.startswith('http'):
|
||||||
|
host = 'http://%s' % self.host
|
||||||
|
if self.port != 80 and self.port is not None:
|
||||||
|
host = '%s:%s' % (host, str(self.port))
|
||||||
|
return host
|
||||||
|
|
||||||
|
def get_post_kwargs(self, payload_input):
|
||||||
|
if self.message_type == 'splunk':
|
||||||
|
# Splunk needs data nested under key "event"
|
||||||
|
if not isinstance(payload_input, dict):
|
||||||
|
payload_input = json.loads(payload_input)
|
||||||
|
payload_input = {'event': payload_input}
|
||||||
|
if isinstance(payload_input, dict):
|
||||||
|
payload_str = json.dumps(payload_input)
|
||||||
|
else:
|
||||||
|
payload_str = payload_input
|
||||||
|
return dict(data=payload_str, background_callback=unused_callback)
|
||||||
|
|
||||||
def emit(self, record):
|
def emit(self, record):
|
||||||
if (self.host == '' or self.enabled_loggers is None or
|
if (self.host == '' or self.enabled_loggers is None or
|
||||||
record.name.split('.')[-1] not in self.enabled_loggers):
|
record.name.split('.')[-1] not in self.enabled_loggers):
|
||||||
return
|
return
|
||||||
try:
|
try:
|
||||||
payload = self.format(record)
|
payload = self.format(record)
|
||||||
host = self.host
|
host = self.get_http_host()
|
||||||
if not host.startswith('http'):
|
|
||||||
host = 'http://%s' % self.host
|
|
||||||
if self.port != 80 and self.port is not None:
|
|
||||||
host = '%s:%s' % (host, str(self.port))
|
|
||||||
|
|
||||||
# Special action for System Tracking, queue up multiple log messages
|
# Special action for System Tracking, queue up multiple log messages
|
||||||
if self.indv_facts:
|
if self.indv_facts:
|
||||||
@@ -133,127 +112,12 @@ class HTTPSHandler(logging.Handler):
|
|||||||
for key in facts_dict:
|
for key in facts_dict:
|
||||||
fact_payload = copy(payload_data)
|
fact_payload = copy(payload_data)
|
||||||
fact_payload.update(facts_dict[key])
|
fact_payload.update(facts_dict[key])
|
||||||
kwargs = dict(data=json.dumps(fact_payload), background_callback=unused_callback)
|
self.session.post(host, **self.get_post_kwargs(fact_payload))
|
||||||
self.session.post(host, **kwargs)
|
|
||||||
return
|
return
|
||||||
|
|
||||||
kwargs = dict(data=payload, background_callback=unused_callback)
|
self.session.post(host, **self.get_post_kwargs(payload))
|
||||||
# # splunk doesn't take "@" in the keys
|
|
||||||
# if self.message_type == 'splunk':
|
|
||||||
# payload_dict = json.loads(payload)
|
|
||||||
# new_dict = {}
|
|
||||||
# for key in payload_dict:
|
|
||||||
# new_key = key.replace('@', '')
|
|
||||||
# new_dict[new_key] = payload_dict[key]
|
|
||||||
# new_payload = json.dumps(new_dict)
|
|
||||||
# kwargs['data'] = json.dumps(new_dict)
|
|
||||||
self.session.post(host, **kwargs)
|
|
||||||
except (KeyboardInterrupt, SystemExit):
|
except (KeyboardInterrupt, SystemExit):
|
||||||
raise
|
raise
|
||||||
except:
|
except:
|
||||||
self.handleError(record)
|
self.handleError(record)
|
||||||
|
|
||||||
|
|
||||||
# splunk
|
|
||||||
# https://github.com/andresriancho/splunk-logger
|
|
||||||
|
|
||||||
class SplunkLogger(logging.Handler):
|
|
||||||
"""
|
|
||||||
A class to send messages to splunk storm using their API
|
|
||||||
"""
|
|
||||||
# Required format for splunk storm
|
|
||||||
INPUT_URL_FMT = 'https://%s/1/inputs/http'
|
|
||||||
|
|
||||||
def __init__(self, access_token=None, project_id=None, api_domain=None):
|
|
||||||
logging.Handler.__init__(self)
|
|
||||||
|
|
||||||
self._set_auth(access_token, project_id, api_domain)
|
|
||||||
self.url = self.INPUT_URL_FMT % self.api_domain
|
|
||||||
|
|
||||||
self._set_url_opener()
|
|
||||||
|
|
||||||
# Handle errors in authentication
|
|
||||||
self._auth_failed = False
|
|
||||||
|
|
||||||
def _set_auth(self, access_token, project_id, api_domain):
|
|
||||||
# The access token and project id passed as parameter override the ones
|
|
||||||
# configured in the .splunk_logger file.
|
|
||||||
if access_token is not None and project_id is not None and api_domain is not None:
|
|
||||||
self.project_id = project_id
|
|
||||||
self.access_token = access_token
|
|
||||||
self.api_domain = api_domain
|
|
||||||
|
|
||||||
else:
|
|
||||||
# Try to get the credentials form the configuration file
|
|
||||||
self.project_id, self.access_token, self.api_domain = parse_config_file()
|
|
||||||
|
|
||||||
if self.project_id is None or self.access_token is None or self.api_domain is None:
|
|
||||||
# Try to get the credentials form the environment variables
|
|
||||||
self.project_id, self.access_token, self.api_domain = get_config_from_env()
|
|
||||||
|
|
||||||
if self.access_token is None or self.project_id is None:
|
|
||||||
raise ValueError('Access token, project id and API endpoint domain'
|
|
||||||
' need to be set.')
|
|
||||||
|
|
||||||
def _set_url_opener(self):
|
|
||||||
# We disable the logging of the requests module to avoid some infinite
|
|
||||||
# recursion errors that might appear.
|
|
||||||
requests_log = logging.getLogger("requests")
|
|
||||||
requests_log.setLevel(logging.CRITICAL)
|
|
||||||
|
|
||||||
self.session = requests.Session()
|
|
||||||
self.session.auth = ('x', self.access_token)
|
|
||||||
self.session.headers.update({'Content-Encoding': 'gzip'})
|
|
||||||
|
|
||||||
def usesTime(self):
|
|
||||||
return False
|
|
||||||
|
|
||||||
def _compress(self, input_str):
|
|
||||||
"""
|
|
||||||
Compress the log message in order to send less bytes to the wire.
|
|
||||||
"""
|
|
||||||
compressed_bits = cStringIO.StringIO()
|
|
||||||
|
|
||||||
f = gzip.GzipFile(fileobj=compressed_bits, mode='wb')
|
|
||||||
f.write(input_str)
|
|
||||||
f.close()
|
|
||||||
|
|
||||||
return compressed_bits.getvalue()
|
|
||||||
|
|
||||||
def emit(self, record):
|
|
||||||
|
|
||||||
if self._auth_failed:
|
|
||||||
# Don't send anything else once a 401 was returned
|
|
||||||
return
|
|
||||||
|
|
||||||
try:
|
|
||||||
response = self._send_to_splunk(record)
|
|
||||||
except (KeyboardInterrupt, SystemExit):
|
|
||||||
raise
|
|
||||||
except:
|
|
||||||
# All errors end here.
|
|
||||||
self.handleError(record)
|
|
||||||
else:
|
|
||||||
if response.status_code == 401:
|
|
||||||
self._auth_failed = True
|
|
||||||
|
|
||||||
def _send_to_splunk(self, record):
|
|
||||||
# http://docs.splunk.com/Documentation/Storm/latest/User/Sourcesandsourcetypes
|
|
||||||
sourcetype = 'json_no_timestamp'
|
|
||||||
|
|
||||||
host = socket.gethostname()
|
|
||||||
|
|
||||||
event_dict = {'data': self.format(record),
|
|
||||||
'level': record.levelname,
|
|
||||||
'module': record.module,
|
|
||||||
'line': record.lineno}
|
|
||||||
event = json.dumps(event_dict)
|
|
||||||
event = self._compress(event)
|
|
||||||
|
|
||||||
params = {'index': self.project_id,
|
|
||||||
'sourcetype': sourcetype,
|
|
||||||
'host': host}
|
|
||||||
|
|
||||||
url = '%s?%s' % (self.url, urllib.urlencode(params))
|
|
||||||
return self.session.post(url, data=event)
|
|
||||||
|
|
||||||
|
|||||||
Reference in New Issue
Block a user