There is a widely used stack of technologies for parsing, collecting and analysing logs - ELK Stack. It has very functional web interface, search cluster and a log transformation tool. Very cool, but:
- It's Java with well-known requirements for memory and CPUs
- It's ElasticSearch with it's requirements for disk space
- It's Logstash which suddenly stops processing logs in some conditions.
- It's Kibana with very cool RICH interface which looses on all counts to
grep
andless
in a task of log reading and searching.
ALCO is a simple ELK analog which primary aim is to provide a online replacement for grep
and less
. Main features are:
- Django application for incident analysis in distributed systems
- schemeless full-text index with filtering and searching
- configurable log collection and rotation from RabbitMQ messaging server
- not a all-purpose monster
Let's trace log message path from some distributed system to ALCO web interface.
- Python-based project calls
logger.debug()
method with text 'hello world' - At startup time Logcollect library automatically configures python logging (or even Django and Celery one's) to send log messages to RabbitMQ server in JSON format readable both with ELK and ALCO projects.
- ALCO log collector binds a queue to RabbitMQ exchange and processes messages in a batch.
- It uses Redis to collect unique values for filterable fields and SphinxSearch to store messages in a realtime index.
- When a message is inserted to sphinxsearch, it contains indexed
message
field, timestamp information and schemeless JSON field namedjs
with all log record attributes sent by python log. - Django-based web interface provides API and client-side app for searching collected logs online.
- Python 2.7 or 3.3+
- Logcollect for python projects which logs are collected
- RabbitMQ server for distributed log collection
- SphinxSearch server 2.3 or later for log storage
- Redis for SphinxSearch docid management and field values storage
- django-sphinxsearch as a database backend for
Django>=1.8
(will be available from PyPI)
-
You need to configure logcollect in analyzed projects (see README). If RabbitMQ admin interface shows non-zero message flow in
logstash
exchange - "It works" :-) -
Install alco and it's requirements from PyPi
pip install alco
-
Next, create django project, add
sphinxsearch
database connection and configuresettings.py
to enable alco applications# For SphinxRouter SPHINX_DATABASE_NAME = 'sphinx' DATABASES[SPHINX_DATABASE_NAME] = { 'ENGINE': 'sphinxsearch.backend.sphinx', 'HOST': '127.0.0.1', 'PORT': 9306, } } # Auto routing log models to SphinxSearch database DATABASE_ROUTERS = ( 'sphinxsearch.routers.SphinxRouter', ) INSTALLED_APPS += [ 'rest_framework', # for API to work 'alco.collector', 'alco.grep' ] ROOT_URLCONF = 'alco.urls'
-
Configure ALCO resources in
settings.py
:ALCO_SETTINGS = { # log messaging server 'RABBITMQ': { 'host': '127.0.0.1', 'userid': 'guest', 'password': 'guest', 'virtual_host': '/' }, # redis server 'REDIS': { 'host': '127.0.0.1', 'db': 0 }, # url for fetching sphinx.conf dynamically 'SPHINX_CONF_URL': 'http://127.0.0.1:8000/collector/sphinx.conf', # name of django.db.connection for SphinxSearch 'SPHINX_DATABASE_NAME': 'sphinx', # number of results in log view API 'LOG_PAGE_SIZE': 100 } # override defaults for sphinx.conf template ALCO_SPHINX_CONF = { # local index definition defaults override 'index': { 'min_word_len': 8 }, # searchd section defaults override 'searchd': { 'dist_threads': 8 } }
-
Run
syncdb
or bettermigrate
management command to create database tables. -
Run webserver and create a LoggerIndex from django admin.
-
Created directories for sphinxsearch:
/var/log/sphinx/ /var/run/sphinx/ /data/sphinx/
-
Next, configure sphinxsearch to use generated config:
searchd -c sphinx_conf.py
sphinx_conf.py
is a simple script that importsalco.sphinx_conf
module which fetches generatedsphinx.conf
from alco http api and created directories for SphinxSearch indices:#!/data/alco/virtualenv/bin/python # coding: utf-8 import os os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'settings') from alco import sphinx_conf
-
Run log collectors:
python manage.py start_collectors --no-daemon
If it shows number of collected messages periodically - then log collecting is set up correctly.
-
Configure system services to start subsystems automatically:
- nginx or apache http server
- django uwsgi backend
- alco collectors (
start_collectors
management command) - sphinxsearch, redis, default database for Django
-
Open
http://127.0.0.1:8000/grep/<logger_name>/
to read and search logs online.
We successfully configured SphinxSearch to use python from virtualenv
, adding some environment variables to start script (i.e. FreeBSD rc.d script):
sphinxsearch_prestart ()
{
# nobody user has no HOME
export PYTHON_EGG_CACHE=/tmp/.python-eggs
# python path for virtualenv interpreter should be redeclared
export PYTHONPATH=${venv_path}/lib/python3.4/:${venv_path}/lib/python3.4/site-packages/
. "${virtualenv_path}/bin/activate" || err 1 "Virtualenv is not found"
echo "Virtualenv ${virtualenv_path} activated: `which python`"
}
In this case shebang for sphinx_conf.py
must point virtualenv's python interpreter.
For now ALCO stack is tested in preproduction environment in our company and is actively developed. There are no reasons to say that it's not ready for production usage.