A HAProxy statistics collection program
Contents
haproxystats is a statistics collector for HAProxy load balancer which processes various statistics and pushes them to graphing systems (Graphite). It is designed to satisfy the following requirements:
- Fast and configurable processing of HAProxy statistics
- Perform aggregation when HAProxy runs in multiprocess (nbproc > 1)
- Pull statistics at very low intervals (10secs)
- Flexible dispatching of statistics to different systems (Graphite, kafka)
The main design characteristic is the split between pulling the statistics and processing them. This provides the ability to pull data as frequently as possible without worrying about the impact on processing time. It also reduces the risk of losing data in case of trouble during the processing phase.
It runs locally on each load balancer node, offering a decentralized setup for the processing phase, but it can be easily extended in the future to have a centralized setup for the processing phase. In that centralized setup it will be possible to perform aggregation on a cluster level as well. Until then users can deploy carbon-c-relay for aggregation.
Because of this design haproxystats comes with two programs: haproxystats-pull and haproxystats-process. The former pulls statistics from HAProxy via stats socket and it uses the asyncio framework from Python to achieve high concurrency and low footprint. The latter processes the statistics and pushes them to various destinations. It utilizes Pandas for data analysis and the multiprocess framework from Python.
haproxystats requires Python 3.4, docopt and Pandas to be available in the system.
haproxystats-pull sends info and stat commands to all haproxy processes in order to collect statistics for the daemon and for all frontends/backends/servers. Data returned from each process and for each command is stored in individual files which are saved under one directory. The time (seconds since the epoch) of retrieval is used to name that directory. haproxystats-process watches for changes on the parent directory and when a directory is created it adds its full path to the queue. Multiple workers pick up items (directories) from the queue and process statistics from those directories.
haproxystats-pull leverages the asyncio framework from Python by utilizing coroutines to multiplex I/O access over several stats socket, which are simple UNIX and TCP sockets.
The actual task of storing the data to the file system is off-loaded to a very light pool of threads in order to avoid blocking the coroutines during the disk IO phase.
haproxystats-pull manages the incoming directory and makes sure directories are created with correct names. It also suspends the collection when the number of directories under the incoming directory exceeds a threshold. This avoids filling up the disk when haproxystats-process is unavailable for sometime. This an example of directory structure:
incoming
├── 1457298067
│ ├── admin1.sock_info
│ ├── admin1.sock_stat
│ ├── admin2.sock_info
│ ├── admin2.sock_stat
│ ├── admin3.sock_info
│ ├── admin3.sock_stat
│ ├── admin4.sock_info
│ └── admin4.sock_stat
└── 1457298072
├── admin1.sock_info
├── admin1.sock_stat
├── admin2.sock_info
├── admin2.sock_stat
├── admin3.sock_info
├── admin3.sock_stat
├── admin4.sock_info
└── admin4.sock_stat
haproxystats-process is a multiprocess program. The parent process uses the Linux kernel's inotify API to watch for changes in incoming directory.
It receives an event when a directory is either created or moved in incoming directory. The event contains the absolute path name of that directory. It maintains an internal queue in which it puts directory names. Multiple child processes pick directory names from the queue and process the data.
Its worker dispatches statistics to various destinations. The directories are removed from incoming directory when all statistics are successfully processed.
When haproxystats-process starts it scans the incoming directory for new directories and processes them instantly, so you don't lose statistics if haproxystats-process is unavailable for sometime.
haproxystats-process currently supports 2 different dispatchers.
- Graphite
Pushes statistics to a Graphite system via a local or remote carbon-relay. The recommended method is to use carbon-c-relay. It is very fast and capable of handling millions of metrics per second. This dispatcher utilizes an internal queue to store metrics which are failed to be sent to Graphite.
An example of graphite namespace:
<loadbalancers>.<lb-01>.haproxy.frontend.<frontendname>. <loadbalancers>.<lb-01>.haproxy.backend.<backendname>. <loadbalancers>.<lb-01>.haproxy.backend.<backendname>.server.<servername> <loadbalancers>.<lb-01>.haproxy.server.<servername>. <loadbalancers>.<lb-01>.haproxy.daemon. <loadbalancers>.<lb-01>.haproxy.haproxystats.<metric names>.
- local-store
Stores statistics in the local disk. Use it only for debugging purposes.
In addition the statistics that are exposed by HAProxy, haproxystats provides the following statistics.
HAProxy exposes Idle_pct and haproxystats-process converts it to CPU utilization without removing Idle_pct metric. This avoids the usage of scale(-1) and offset(100) functions on graphite:
CpuUsagePct CPU utilization in percentage
The following metrics are calculated only when HAProxy is configured with more than 1 processes (nbproc > 1):
25PercentileCpuUsagePct 25th percentile of CpuUsagePct across all processes 50PercentileCpuUsagePct 50th percentile -//- 75PercentileCpuUsagePct 75th percentile -//- 95PercentileCpuUsagePct 95th percentile -//- 99PercentileCpuUsagePct 99th percentile -//- StdCpuUsagePct standard deviation -//-
The incoming directory together with the inotify API provides a simple queueing system which is used as a communication channel between haproxystats-pull and haproxystats-process programs.
There isn't any feedback mechanism in place, thus haproxystats-pull monitors the number of directories before it pulls data from HAProxy and suspends its job when the number of directories exceeds a threshold.
See queue-size parameter of pull section.
haproxystats provides statistics for the time it takes to process, calculate and send HAProxy metrics. By default provides the following list of metric names with values in seconds:
loadbalancers.lb-01.haproxy.haproxystats.WallClockTimeHAProxy loadbalancers.lb-01.haproxy.haproxystats.WallClockTimeFrontends loadbalancers.lb-01.haproxy.haproxystats.WallClockTimeBackends loadbalancers.lb-01.haproxy.haproxystats.WallClockTimeServers loadbalancers.lb-01.haproxy.haproxystats.WallClockTimeAllStats
It also provides the number of metrics which are send to graphite:
loadbalancers.lb-01.haproxy.haproxystats.MetricsHAProxy loadbalancers.lb-01.haproxy.haproxystats.MetricsFrontend loadbalancers.lb-01.haproxy.haproxystats.MetricsBackend loadbalancers.lb-01.haproxy.haproxystats.MetricsServer
haproxystats uses the popular INI format for its configuration file. This is an example configuration file (/etc/haproxystats.conf):
[DEFAULT] loglevel = info retries = 2 timeout = 1 interval = 2 [paths] base-dir = /var/lib/haproxystats [pull] loglevel = info socket-dir = /run/haproxy retries = 1 timeout = 0.1 interval = 0.5 pull-timeout = 2 pull-interval = 10 dst-dir = ${paths:base-dir}/incoming tmp-dst-dir = ${paths:base-dir}/incoming.tmp workers = 8 queue-size = 360 [process] src-dir = ${paths:base-dir}/incoming workers = 4 per-process-metrics = false [graphite] server = 127.0.0.1 port = 3002 retries = 3 interval = 1.8 connect-timeout = 1.0 write-timeout = 1.0 delay = 10 backoff = 2 namespace = loadbalancers prefix-hostname = true fqdn = true queue-size = 1000000 #[local-store] #dir = ${paths:base-dir}/local-store
All the above settings are optional as haproxystats comes with default values for all of them. Thus, both programs can be started without supplying any configuration.
Settings in this section can be overwritten in other sections.
- loglevel Defaults to info
Log level to use, possible values are: debug, info, warning, error, critical
- retries Defaults to 2
Number of times to retry a connection after a failure. Used by haproxystats-pull and haproxystats-process when they open a connection to a UNIX/TCP socket and Graphite respectively.
- timeout Defaults to 1 (seconds)
Time to wait for establishing a connection. Used by haproxystats-pull and haproxystats-process when they open a connection to a UNIX/TCP socket and Graphite respectively.
- interval Defaults to 2
Time to wait before trying to open a connection. Used by haproxystats-pull and haproxystats-process when they retry a connection to a UNIX/TCP socket and Graphite respectively.
- base-dir Defaults to /var/lib/haproxystats
The directory to use as the base of the directory structure.
- socket-dir Unset by default
A directory with HAProxy socket files.
- servers Unset by default
A list of servers to pull statistics from. You define a server by passing a URL, here some examples:
tcp://127.0.0.1:5555 tcp://foo.bar.com:4444 tcp://[fe80::3f2f:46b3:ef0c:a420]:4444 unix:///run/haproxy.sock
Only TCP and UNIX schemes are supported and the port for TCP servers must be set. For UNIX scheme you can only pass a file and not a directory, but socket-dir option can be set as well, so you can use a directory and UNIX socket files at the same time. You can use comma as separator to pass multiple servers:
servers = unix:///run/haproxy.sock,tcp://127.0.0.1:555,tcp://127.0.0.1:556
- buffer-limit Defaults to 6291456 (bytes)
At most size bytes are read and returned from the sockets. Setting too low and it will slow down the retrieval of statistics. Only values greater than or equal to 1 are accepted.
- retries Defaults to 1
Number of times to reconnect to UNIX/TCP socket after a failure.
- timeout Defaults to 0.1 (seconds)
Time to wait for establishing a connection to UNIX/TCP socket. There is no need to set it higher than few ms as haproxy accepts a connection within 1-2ms.
- interval Defaults to 0.5 (seconds)
Time to wait before trying to reconnect to UNIX/TCP socket after a failure. Tune it based on the duration of the reload process of haproxy. haproxy reloads within few ms but in some environments with hundreds different SSL certificates it can take a bit more.
- pull-interval Defaults to 10 (seconds)
How often to pull statistics from HAProxy. A value of 1 second can overload the haproxy processes in environments with thousands backends/servers.
- pull-timeout Defaults to 2 (seconds)
Total time to wait for the pull process to finish. Should be always less than pull-interval.
- dst-dir Defaults /var/lib/haproxystats/incoming
A directory to store statistics retrieved by HAProxy.
- tmp-dst-dir Defaults /var/lib/haproxystats/incoming.tmp
A directory to use as temporary storage location before directories are moved to dst-dir. haproxystats-pull stores statistics for each process under that directory and only when data from all haproxy processes are successfully retrieved they are moved to dst-dir. Make sure dst-dir and tmp-dst-dir are on the same file system, so the move of the directories become a rename which is a quick and atomic operation.
- workers Defaults to 8
Number of threads to use for writing statistics to disk. These are very light threads and don't consume a lot of resources. Shouldn't be set higher than the number of haproxy processes.
- queue-size Defaults to 360
Suspend the pulling of statistics when the number of directories in dst-dir exceeds this limit.
- src-dir Defaults /var/lib/haproxystats/incoming
A directory to watch for changes. It should point to the same directory as the dst-dir option from pull section.
- workers Defaults to 4
Number of workers to use for processing statistics. These are real processes which can consume a fair bit of CPU.
- frontend-metrics Unset by default
A list of frontend metric names separated by space to process. By default all statistics are processed and this overwrites the default selection.
haproxystats-process emits an error and refuses to start if metrics aren't valid HAProxy metrics. Check the list of valid metrics in Chapter 9.1 of management documentation of HAProxy.
- backend-metrics Unset by default
A list of backend metric names separated by space to process. By default all statistics are processed and this overwrites the default selection.
haproxystats-process emits an error and refuses to start if metrics aren't valid HAProxy metrics. Check the list of valid metrics in Chapter 9.1 of management documentation of HAProxy.
- server-metrics Unset by default
A list of server metric names separated by space to process. By default all statistics are processed and this overwrites the default selection.
haproxystats-process emits an error and refuses to start if metrics aren't valid HAProxy metrics. Check the list of valid metrics in Chapter 9.1 of management documentation of HAProxy.
- aggr-server-metrics Defaults to false
Aggregates server's statistics across all backends.
- exclude-frontends Unset by default
A file which contains one frontend name per line for which processing is skipped.
- exclude-backends Unset by default
A file which contains one backend name per line for which processing is skipped.
- per-process-metrics Defaults to false
HAProxy daemon provides statistics and by default haproxystat-process aggregates those statistics when HAProxy runs in multiprocess mode (nbproc > 1).
Set this to true to get those statistics also per process as well. This is quite useful for monitoring purposes where someone wants to monitor sessions per process in order to see if traffic is evenly distributed to all processes by the kernel.
It is also useful in setups where configuration for frontends and backends is unevenly spread across all processes, for instance processes 1-4 manage SSL frontends and processes 5-7 manage noSSL frontends.
This adds another path in Graphite under haproxy space:
loadbalancers.lb-01.haproxy.daemon.process.<process_num>.<metric>
- calculate-percentages Defaults to false
Calculates percentages for a selection of metrics for HAProxy daemon. When per-process-metrics is set to true the calculation happens also per HAProxy process. This adds the following metric names:
ConnPercentage ConnRatePercentage SslRatePercentage SslConnPercentage
Those metrics can be used for alerting when the current usage on connections is very close the configured limit.
- liveness-check-interval Defaults to 10 (seconds)
How often to check if all workers are alive and trigger a termination if at least one is dead.
This dispatcher is enabled by default and it can't be disabled.
- server Defaults to 127.0.0.1
Graphite server to connect to.
- port Defaults to 3002
Graphite port to connect to.
- retries Defaults to 3
Number of times to reconnect to Graphite after a failure.
- interval Defaults to 1.8 (seconds)
Time to wait before trying to reconnect to Graphite after a failure.
- connect-timeout Defaults to 1 (seconds)
Time to wait for establishing a connection to Graphite relay.
- write-timeout Defaults to 1 (seconds)
Time to wait on sending data to Graphite relay.
- delay Defaults to 10 (seconds)
How long to wait before trying to connect again after number of retries has exceeded the threshold set in retries. During the delay period metrics are stored in the queue of the dispatcher, see queue-size.
- backoff Defaults to 2
A simple exponential backoff to apply for each retry.
- namespace Defaults to loadbalancers
A top level graphite namespace.
- prefix-hostname Defaults to true
Insert the hostname of the load balancer in the Graphite namespace, example:
loadbalancers.lb-01.haproxy.
- fqdn Defaults to true
Use FQDN or short name in the graphite namespace
- queue-size Defaults to 1000000
haproxystats-process uses a queue to store metrics which failed to be sent due to a connection error/timeout. This is a First In First Out queueing system. When the queue reaches the limit, the oldest items are removed to free space.
- group-namespace Unset by default.
group graphite metrics by patterns. When a frontend, backend or server matches a given pattern, the metric will be prefixed by this namespace, plus a configurable group name which must be specified in the frontend-groups, backend-groups or server-groups sections. These sections consist of group names and their corresponding regular expression that will be matched against frontend, backend or server names (depending on the section).
For example:
Let's assume our metrics look something like:
loadbalancers.lb-01.haproxy.frontend.foo-001.<metric> loadbalancers.lb-01.haproxy.frontend.foo-002.<metric> ... loadbalancers.lb-01.haproxy.frontend.bar-001.<metric> loadbalancers.lb-01.haproxy.frontend.bar-002.<metric> ...
And we want them to be grouped to like this:
loadbalancers.lb-01.haproxy.flavor.abc.frontend.foo-001.<metric> loadbalancers.lb-01.haproxy.flavor.abc.frontend.foo-002.<metric> ... loadbalancers.lb-01.haproxy.flavor.xyz.frontend.bar-001.<metric> loadbalancers.lb-01.haproxy.flavor.xyz.frontend.bar-002.<metric> ...
The configuration should contain these settings:
[graphite] group-namespace = flavor [frontend-groups] abc = ^foo- xyz = ^bar-
Note that if the group-namespace setting is specified, then at least one of frontend-groups, backend-groups or server-groups sections must be specified as well.
Also note that if frontend, backend or server names contain dots, these will be converted to underscores for graphite -- because dots are graphite's namespace separator. The patterns will have to take this into account.
- group-namespace-double-writes Unset by default.
Boolean; required only if group-namespace is specified. If True, send to graphite the original metric as well as the grouped metrics. If False, send only the grouped metrics. (See group-namespace.)
Specify the patterns to match against frontend, backend and/or server names, to group graphite metrics and give them a variable prefix. See group-namespace.
These sections are optional, unless group-namespace is set.
This dispatcher isn't enabled by default.
The primarily use of local-store dispatcher is to debug/troubleshoot possible problems with the processing or/and with Graphite. There isn't any clean-up process in place, thus you need remove the files after they are created. Don't leave it enabled for more than 1 hour as it can easily fill up the disk in environments with hundreds frontends/backends and thousands servers.
- dir Defaults to /var/lib/haproxystats/local-store
A directory to stores statistics after they have been processed. The current format is compatible with Graphite.
haproxystats-pull and haproxystats-process are simple programs which are not daemonized and they output logging messages to stdout. This is by design as it simplifies the code. The daemonization and logging is off-loaded to systemd which has everything we need for that job.
Under contrib/systend directory there are service files for both programs. These are functional systemd Unit files which are used in production.
The order in which these 2 programs start doesn't matter and there isn't any soft or hard dependency between them.
Furthermore, these programs don't need to run as root. It highly recommended to create a dedicated user to run them. You need to add that user to the group of haproxy and adjust socket configuration of haproxy to allow write for the group, see below an example configuration:
stats socket /run/haproxy/sock1 user haproxy group haproxy mode 660 level admin process 1 stats socket /run/haproxy/sock2 user haproxy group haproxy mode 660 level admin process 2 stats socket /run/haproxy/sock3 user haproxy group haproxy mode 660 level admin process 3
systemd Unit files use haproxystats user which has to be created prior running haproxystats programs.
In an effort to reduce the loss of statistics both programs support graceful shutdown. When SIGHUP or SIGTERM signals are sent they perform a clean exit. When a signal is sent to haproxystats-process it may take some time for the program to exit, as it waits for all workers to empty the queue.
A puppet module is available under contrib directory which provides classes for configuring both programs.
Because haproxystats-process is CPU bound program, CPU Affinity is configured using systemd. By default it pins the workers to the last CPUs.
You should take care of pinning haproxy processes to other CPUs in order to avoid haproxystats-process stealing CPU cycles from haproxy. In production servers you usually pin the first 80% of CPUs to haproxy processes and you leave the rest of CPUs for other processes. The default template of puppet module enforces this logic.
haproxystats-pull is a single threaded program which doesn't use a lot of CPU cycles and by default is assigned to the last CPU.
A Ansible playbook is available under contrib directory. For installation instruction of the playbook please read Installation chapter of this document.
Several nagios checks are provided for monitoring purposes, they can be found under contrib/nagios directory.
- check_haproxystats_process_number_of_procs.sh
Monitor the number of processes of haproxystats-process program. Systemd monitors only the parent process and this check helps to detect cases where some worker(s) die unexpectedly
- check_haproxystats_process.sh
A wrapper around systemctl tool to detect a dead parent process.
- check_haproxystats_pull.sh
A wrapper around systemctl tool to a check if haproxystats-pull is running.
- check_haproxystats_queue_size.py
Checks the size of the incoming directory queue which is consumed by haproxystats-process and alert when exceeds a threshold.
haproxystats-pull -f ./haproxystats.conf
haproxystats-process -f ./haproxystats.conf
Usage:
% haproxystats-pull -h Pulls statistics from HAProxy daemon over UNIX socket(s) Usage: haproxystats-pull [-f <file>] [-p | -P] Options: -f, --file <file> configuration file with settings [default: /etc/haproxystats.conf] -p, --print show default settings -P, --print-conf show configuration -h, --help show this screen -v, --version show version % haproxystats-process -h Processes statistics from HAProxy and pushes them to Graphite Usage: haproxystats-process [-f <file>] [-d <dir>] [-p | -P] Options: -f, --file <file> configuration file with settings [default: /etc/haproxystats.conf] -d, --dir <dir> directory with additional configuration files -p, --print show default settings -P, --print-conf show configuration -h, --help show this screen -v, --version show version
I would love to hear what other people think about haproxystats and provide feedback. Please post your comments, bug reports and wishes on my issues page.
Install HAProxy:
% sudo apt-get install haproxy
Use a basic HAProxy configuration in multiprocess mode:
global log 127.0.0.1 len 2048 local2 chroot /var/lib/haproxy stats socket /run/haproxy/admin1.sock mode 666 level admin process 1 stats socket /run/haproxy/admin2.sock mode 666 level admin process 2 stats socket /run/haproxy/admin3.sock mode 666 level admin process 3 stats socket /run/haproxy/admin4.sock mode 666 level admin process 4 # allow read/write access to anyone----------^ stats timeout 30s user haproxy group haproxy daemon nbproc 4 cpu-map 1 0 cpu-map 2 1 cpu-map 3 1 cpu-map 4 0 defaults log global mode http timeout connect 5000 timeout client 50000 timeout server 50000 frontend frontend_proc1 bind 0.0.0.0:81 process 1 default_backend backend_proc1 frontend frontend_proc2 bind 0.0.0.0:82 process 2 default_backend backend_proc1 frontend frontend1_proc34 bind :83 process 3 bind :83 process 4 default_backend backend1_proc34 backend backend_proc1 bind-process 1 default-server inter 1000s option httpchk GET / HTTP/1.1\r\nHost:\ .com\r\nUser-Agent:\ HAProxy server member1_proc1 10.189.224.169:80 weight 100 check fall 2 rise 3 server member2_proc1 10.196.70.109:80 weight 100 check fall 2 rise 3 server bck_all_srv1 10.196.70.109:88 weight 100 check fall 2 rise 3 backend backend1_proc34 bind-process 3,4 default-server inter 1000s option httpchk GET / HTTP/1.1\r\nHost:\ .com\r\nUser-Agent:\ HAProxy server bck1_proc34_srv1 10.196.70.109:80 check fall 2 inter 5s rise 3 server bck1_proc34_srv2 10.196.70.109:80 check fall 2 inter 5s rise 3 server bck_all_srv1 10.196.70.109:80 check fall 2 inter 5s rise 3 backend backend_proc2 bind-process 2 default-server inter 1000s option httpchk GET / HTTP/1.1\r\nHost:\ .com\r\nUser-Agent:\ HAProxy server bck_proc2_srv1_proc2 127.0.0.1:8001 check fall 2 inter 5s rise 3 server bck_proc2_srv2_proc2 127.0.0.1:8002 check fall 2 inter 5s rise 3 server bck_proc2_srv3_proc2 127.0.0.1:8003 check fall 2 inter 5s rise 3 server bck_proc2_srv4_proc2 127.0.0.1:8004 check fall 2 inter 5s rise 3
Start HAProxy and check it is up:
sudo systemctl start haproxy.service;systemctl status -l haproxy.service
Create a python virtual environment using virtualenvwrapper tool:
mkvirtualenv --python=`which python3` haproxystats-dev
Do not exit the haproxystats-dev virtual environment.
Clone the project, if you are planning to contribute then you should fork it on GitHub and clone that project instead:
mkdir ~/repo;cd ~/repo git clone https://github.com/unixsurfer/haproxystats
Install necessary libraries:
cd haproxystats pip install -U pbr setuptools pip install -r ./requirements.txt
Start a TCP server which acts a Graphite relay and listens on 127.0.0.1:39991:
python3 ./contrib/tcp_server.py
Install haproxystats:
python setup.py install
Create necessary directory structure:
mkdir -p ./var/var/lib/haproxystats mkdir -p ./var/etc mkdir -p ./var/etc/haproxystats.d
Adjust the following configuration and save it in ./var/etc/haproxystats.conf:
[DEFAULT] loglevel = debug retries = 2 timeout = 1 interval = 2 [paths] base-dir = /home/<username>/repo/haproxystats/var/var/lib/haproxystats [pull] socket-dir = /run/haproxy retries = 1 timeout = 0.1 interval = 0.5 pull-timeout = 10 pull-interval = 10 dst-dir = ${paths:base-dir}/incoming tmp-dst-dir = ${paths:base-dir}/incoming.tmp workers = 8 [process] src-dir = ${paths:base-dir}/incoming workers = 2 calculate-percentages = true per-process-metrics = true [graphite] server = 127.0.0.1 port = 39991 retries = 3 interval = 0.8 timeout = 0.9 delay = 10 backoff = 2 namespace = loadbalancers prefix_hostname = true fqdn = true queue-size = 1000 [local-store] dir = ${paths:base-dir}/local-store
Start haproxystats-pull and haproxystats-process on 2 different terminals:
haproxystats-pull -f var/etc/haproxystats.conf haproxystats-process -f var/etc/haproxystats.conf
Exit from haproxystats-dev virtual environment:
deactivate
Start hacking and don't forget to make a Pull Request
Use pip:
pip install haproxystats
From Source:
sudo python setup.py install
Build (source) RPMs:
python setup.py clean --all; python setup.py bdist_rpm
Build a source archive for manual installation:
python setup.py sdist
Use Ansible Playbook:
To deploy haproxystats By Ansible Playbook go to contrib/ansible-playbook directory:
cd contrib/ansible-playbook
Then enter your haproxy server IP address in hosts file:
vi hosts
After that set information of your environment in variable file:
vi group_vars/all
Now for run Ansible Playbook use this command:
ansible-playbook -i hosts main-playbook.yml
When Ansible Playbook run successful completely, you can take control haproxystats-pull and haproxystats-process by systemd:
systemctl start haproxystats-pull.service systemctl start haproxystats-process.service
Bump version in haproxystats/__init__.py
Commit above change with:
git commit -av -m'RELEASE 0.1.3 version'
Create a signed tag, pbr will use this for the version number:
git tag -s 0.1.3 -m 'bump release'
Create the source distribution archive (the archive will be placed in the dist directory):
python setup.py sdist
pbr updates ChangeLog file and we want to squeeze this change to the previous commit, thus run:
git commit -av --amend
Move current tag to the last commit:
git tag -fs 0.1.3 -m 'bump release'
Push changes:
git push;git push --tags
Upload to Python Package Index:
twine upload -s -p dist/*
The following people have contributed to project with feedback and code reviews
- Károly Nagy https://github.com/charlesnagy
- Dan Achim https://github.com/danakim
Apache 2.0
This program was originally developed for Booking.com. With approval from Booking.com, the code was generalised and published as Open Source on github, for which the author would like to express his gratitude.
Project website: https://github.com/unixsurfer/haproxystats
Author: Pavlos Parissis <pavlos.parissis@gmail.com>