Skip to main content

A HAProxy statistics collection program

Project description

A HAProxy statistics collection program

Introduction

haproxystats is a statistics collector for HAProxy load balancer which processes various statistics and pushes them to graphing systems (Graphite). It is designed to satisfy the following requirements:

  1. Fast and configurable processing of HAProxy statistics

  2. Perform aggregation when HAProxy runs in multiprocess (nbproc > 1)

  3. Pull statistics at very low intervals (10secs)

  4. Flexible dispatching of statistics to different systems (Graphite, kafka)

The main design characteristic is the split between pulling the statistics and processing them. This provides the ability to pull data as frequently as possible without worrying about the impact on processing time. It also reduces the risk of losing data in case of trouble during the processing phase.

It runs locally on each load balancer node, offering a decentralized setup for the processing phase, but it can be easily extended in the future to have a centralized setup for the processing phase. In that centralized setup it will be possible to perform aggregation on a cluster level as well. Until then users can deploy carbon-c-relay for aggregation.

Because of this design haproxystats comes with two programs: haproxystats-pull and haproxystats-process. The former pulls statistics from HAProxy via stats socket and it uses the asyncio framework from Python to achieve high concurrency and low footprint. The latter processes the statistics and pushes them to various destinations. It utilizes Pandas for data analysis and the multiprocess framework from Python.

haproxystats requires Python 3.4, docopt and Pandas to be available in the system.

How haproxystats works

haproxystats-architecture.png

haproxystats-pull sends info and stat commands to all haproxy processes in order to collect statistics for the daemon and for all frontends/backends/servers. Data returned from each process and for each command is stored in individual files which are saved under one directory. The time (seconds since the epoch) of retrieval is used to name that directory. haproxystats-process watches for changes on the parent directory and when a directory is created it adds its full path to the queue. Multiple workers pick up items (directories) from the queue and process statistics from those directories.

haproxystats-pull

haproxystats-pull leverages the asyncio framework from Python by utilizing coroutines to multiplex I/O access over several stats socket, which are simple UNIX and TCP sockets.

The actual task of storing the data to the file system is off-loaded to a very light pool of threads in order to avoid blocking the coroutines during the disk IO phase.

haproxystats-pull manages the incoming directory and makes sure directories are created with correct names. It also suspends the collection when the number of directories under the incoming directory exceeds a threshold. This avoids filling up the disk when haproxystats-process is unavailable for sometime. This an example of directory structure:

incoming
├── 1457298067
   ├── admin1.sock_info
   ├── admin1.sock_stat
   ├── admin2.sock_info
   ├── admin2.sock_stat
   ├── admin3.sock_info
   ├── admin3.sock_stat
   ├── admin4.sock_info
   └── admin4.sock_stat
└── 1457298072
    ├── admin1.sock_info
    ├── admin1.sock_stat
    ├── admin2.sock_info
    ├── admin2.sock_stat
    ├── admin3.sock_info
    ├── admin3.sock_stat
    ├── admin4.sock_info
    └── admin4.sock_stat

haproxystats-process

haproxystats-process is a multiprocess program. The parent process uses the Linux kernel’s inotify API to watch for changes in incoming directory.

It receives an event when a directory is either created or moved in incoming directory. The event contains the absolute path name of that directory. It maintains an internal queue in which it puts directory names. Multiple child processes pick directory names from the queue and process the data.

Its worker dispatches statistics to various destinations. The directories are removed from incoming directory when all statistics are successfully processed.

When haproxystats-process starts it scans the incoming directory for new directories and processes them instantly, so you don’t lose statistics if haproxystats-process is unavailable for sometime.

Dispatchers

haproxystats-process currently supports 2 different dispatchers.

  1. Graphite

Pushes statistics to a Graphite system via a local or remote carbon-relay. The recommended method is to use carbon-c-relay. It is very fast and capable of handling millions of metrics per second. This dispatcher utilizes an internal queue to store metrics which are failed to be sent to Graphite.

An example of graphite namespace:

<loadbalancers>.<lb-01>.haproxy.frontend.<frontendname>.
<loadbalancers>.<lb-01>.haproxy.backend.<backendname>.
<loadbalancers>.<lb-01>.haproxy.backend.<backendname>.server.<servername>
<loadbalancers>.<lb-01>.haproxy.server.<servername>.
<loadbalancers>.<lb-01>.haproxy.daemon.
<loadbalancers>.<lb-01>.haproxy.haproxystats.<metric names>.
  1. local-store

Stores statistics in the local disk. Use it only for debugging purposes.

Statistics for HAProxy

In addition the statistics that are exposed by HAProxy, haproxystats provides the following statistics.

HAProxy process

HAProxy exposes Idle_pct and haproxystats-process converts it to CPU utilization without removing Idle_pct metric. This avoids the usage of scale(-1) and offset(100) functions on graphite:

CpuUsagePct  CPU utilization in percentage

The following metrics are calculated only when HAProxy is configured with more than 1 processes (nbproc > 1):

25PercentileCpuUsagePct 25th percentile of CpuUsagePct across all processes
50PercentileCpuUsagePct 50th percentile              -//-
75PercentileCpuUsagePct 75th percentile              -//-
95PercentileCpuUsagePct 95th percentile              -//-
99PercentileCpuUsagePct 99th percentile              -//-
StdCpuUsagePct          standard deviation           -//-

Queuing system

The incoming directory together with the inotify API provides a simple queueing system which is used as a communication channel between haproxystats-pull and haproxystats-process programs.

There isn’t any feedback mechanism in place, thus haproxystats-pull monitors the number of directories before it pulls data from HAProxy and suspends its job when the number of directories exceeds a threshold.

See queue-size parameter of pull section.

Statistics for haproxystats

haproxystats provides statistics for the time it takes to process, calculate and send HAProxy metrics. By default provides the following list of metric names with values in seconds:

loadbalancers.lb-01.haproxy.haproxystats.WallClockTimeHAProxy
loadbalancers.lb-01.haproxy.haproxystats.WallClockTimeFrontends
loadbalancers.lb-01.haproxy.haproxystats.WallClockTimeBackends
loadbalancers.lb-01.haproxy.haproxystats.WallClockTimeServers
loadbalancers.lb-01.haproxy.haproxystats.WallClockTimeAllStats

It also provides the number of metrics which are send to graphite:

loadbalancers.lb-01.haproxy.haproxystats.MetricsHAProxy
loadbalancers.lb-01.haproxy.haproxystats.MetricsFrontend
loadbalancers.lb-01.haproxy.haproxystats.MetricsBackend
loadbalancers.lb-01.haproxy.haproxystats.MetricsServer

Configuration

haproxystats uses the popular INI format for its configuration file. This is an example configuration file (/etc/haproxystats.conf):

[DEFAULT]
loglevel = info
retries  = 2
timeout  = 1
interval = 2

[paths]
base-dir = /var/lib/haproxystats

[pull]
loglevel        = info
socket-dir      = /run/haproxy
retries         = 1
timeout         = 0.1
interval        = 0.5
pull-timeout    = 2
pull-interval   = 10
dst-dir         = ${paths:base-dir}/incoming
tmp-dst-dir     = ${paths:base-dir}/incoming.tmp
workers         = 8
queue-size      = 360

[process]
src-dir             = ${paths:base-dir}/incoming
workers             = 4
per-process-metrics = false

[graphite]
server          = 127.0.0.1
port            = 3002
retries         = 3
interval        = 1.8
connect-timeout = 1.0
write-timeout   = 1.0
delay           = 10
backoff         = 2
namespace       = loadbalancers
prefix-hostname = true
fqdn            = true
queue-size      = 1000000

#[local-store]
#dir = ${paths:base-dir}/local-store

All the above settings are optional as haproxystats comes with default values for all of them. Thus, both programs can be started without supplying any configuration.

DEFAULT section

Settings in this section can be overwritten in other sections.

  • loglevel Defaults to info

Log level to use, possible values are: debug, info, warning, error, critical

  • retries Defaults to 2

Number of times to retry a connection after a failure. Used by haproxystats-pull and haproxystats-process when they open a connection to a UNIX/TCP socket and Graphite respectively.

  • timeout Defaults to 1 (seconds)

Time to wait for establishing a connection. Used by haproxystats-pull and haproxystats-process when they open a connection to a UNIX/TCP socket and Graphite respectively.

  • interval Defaults to 2

Time to wait before trying to open a connection. Used by haproxystats-pull and haproxystats-process when they retry a connection to a UNIX/TCP socket and Graphite respectively.

paths section

  • base-dir Defaults to /var/lib/haproxystats

The directory to use as the base of the directory structure.

pull section

  • socket-dir Unset by default

A directory with HAProxy socket files.

  • servers Unset by default

A list of servers to pull statistics from. You define a server by passing a URL, here some examples:

tcp://127.0.0.1:5555
tcp://foo.bar.com:4444
tcp://[fe80::3f2f:46b3:ef0c:a420]:4444
unix:///run/haproxy.sock

Only TCP and UNIX schemes are supported and the port for TCP servers must be set. For UNIX scheme you can only pass a file and not a directory, but socket-dir option can be set as well, so you can use a directory and UNIX socket files at the same time. You can use comma as separator to pass multiple servers:

servers = unix:///run/haproxy.sock,tcp://127.0.0.1:555,tcp://127.0.0.1:556
  • buffer-limit Defaults to 6291456 (bytes)

At most size bytes are read and returned from the sockets. Setting too low and it will slow down the retrieval of statistics. Only values greater than or equal to 1 are accepted.

  • retries Defaults to 1

Number of times to reconnect to UNIX/TCP socket after a failure.

  • timeout Defaults to 0.1 (seconds)

Time to wait for establishing a connection to UNIX/TCP socket. There is no need to set it higher than few ms as haproxy accepts a connection within 1-2ms.

  • interval Defaults to 0.5 (seconds)

Time to wait before trying to reconnect to UNIX/TCP socket after a failure. Tune it based on the duration of the reload process of haproxy. haproxy reloads within few ms but in some environments with hundreds different SSL certificates it can take a bit more.

  • pull-interval Defaults to 10 (seconds)

How often to pull statistics from HAProxy. A value of 1 second can overload the haproxy processes in environments with thousands backends/servers.

  • pull-timeout Defaults to 2 (seconds)

Total time to wait for the pull process to finish. Should be always less than pull-interval.

  • dst-dir Defaults /var/lib/haproxystats/incoming

A directory to store statistics retrieved by HAProxy.

  • tmp-dst-dir Defaults /var/lib/haproxystats/incoming.tmp

A directory to use as temporary storage location before directories are moved to dst-dir. haproxystats-pull stores statistics for each process under that directory and only when data from all haproxy processes are successfully retrieved they are moved to dst-dir. Make sure dst-dir and tmp-dst-dir are on the same file system, so the move of the directories become a rename which is a quick and atomic operation.

  • workers Defaults to 8

Number of threads to use for writing statistics to disk. These are very light threads and don’t consume a lot of resources. Shouldn’t be set higher than the number of haproxy processes.

  • queue-size Defaults to 360

Suspend the pulling of statistics when the number of directories in dst-dir exceeds this limit.

process section

  • src-dir Defaults /var/lib/haproxystats/incoming

A directory to watch for changes. It should point to the same directory as the dst-dir option from pull section.

  • workers Defaults to 4

Number of workers to use for processing statistics. These are real processes which can consume a fair bit of CPU.

  • frontend-metrics Unset by default

A list of frontend metric names separated by space to process. By default all statistics are processed and this overwrites the default selection.

haproxystats-process emits an error and refuses to start if metrics aren’t valid HAProxy metrics. Check the list of valid metrics in Chapter 9.1 of management documentation of HAProxy.

  • backend-metrics Unset by default

A list of backend metric names separated by space to process. By default all statistics are processed and this overwrites the default selection.

haproxystats-process emits an error and refuses to start if metrics aren’t valid HAProxy metrics. Check the list of valid metrics in Chapter 9.1 of management documentation of HAProxy.

  • server-metrics Unset by default

A list of server metric names separated by space to process. By default all statistics are processed and this overwrites the default selection.

haproxystats-process emits an error and refuses to start if metrics aren’t valid HAProxy metrics. Check the list of valid metrics in Chapter 9.1 of management documentation of HAProxy.

  • aggr-server-metrics Defaults to false

Aggregates server’s statistics across all backends.

  • exclude-frontends Unset by default

A file which contains one frontend name per line for which processing is skipped.

  • exclude-backends Unset by default

A file which contains one backend name per line for which processing is skipped.

  • per-process-metrics Defaults to false

HAProxy daemon provides statistics and by default haproxystat-process aggregates those statistics when HAProxy runs in multiprocess mode (nbproc > 1).

Set this to true to get those statistics also per process as well. This is quite useful for monitoring purposes where someone wants to monitor sessions per process in order to see if traffic is evenly distributed to all processes by the kernel.

It is also useful in setups where configuration for frontends and backends is unevenly spread across all processes, for instance processes 1-4 manage SSL frontends and processes 5-7 manage noSSL frontends.

This adds another path in Graphite under haproxy space:

loadbalancers.lb-01.haproxy.daemon.process.<process_num>.<metric>
  • calculate-percentages Defaults to false

Calculates percentages for a selection of metrics for HAProxy daemon. When per-process-metrics is set to true the calculation happens also per HAProxy process. This adds the following metric names:

ConnPercentage
ConnRatePercentage
SslRatePercentage
SslConnPercentage

Those metrics can be used for alerting when the current usage on connections is very close the configured limit.

graphite section

This dispatcher is enabled by default and it can’t be disabled.

  • server Defaults to 127.0.0.1

Graphite server to connect to.

  • port Defaults to 3002

Graphite port to connect to.

  • retries Defaults to 3

Number of times to reconnect to Graphite after a failure.

  • interval Defaults to 1.8 (seconds)

Time to wait before trying to reconnect to Graphite after a failure.

  • connect-timeout Defaults to 1 (seconds)

Time to wait for establishing a connection to Graphite relay.

  • write-timeout Defaults to 1 (seconds)

Time to wait on sending data to Graphite relay.

  • delay Defaults to 10 (seconds)

How long to wait before trying to connect again after number of retries has exceeded the threshold set in retries. During the delay period metrics are stored in the queue of the dispatcher, see queue-size.

  • backoff Defaults to 2

A simple exponential backoff to apply for each retry.

  • namespace Defaults to loadbalancers

A top level graphite namespace.

  • prefix-hostname Defaults to true

Insert the hostname of the load balancer in the Graphite namespace, example:

loadbalancers.lb-01.haproxy.
  • fqdn Defaults to true

Use FQDN or short name in the graphite namespace

  • queue-size Defaults to 1000000

haproxystats-process uses a queue to store metrics which failed to be sent due to a connection error/timeout. This is a First In First Out queueing system. When the queue reaches the limit, the oldest items are removed to free space.

local-store section

This dispatcher isn’t enabled by default.

The primarily use of local-store dispatcher is to debug/troubleshoot possible problems with the processing or/and with Graphite. There isn’t any clean-up process in place, thus you need remove the files after they are created. Don’t leave it enabled for more than 1 hour as it can easily fill up the disk in environments with hundreds frontends/backends and thousands servers.

  • dir Defaults to /var/lib/haproxystats/local-store

A directory to stores statistics after they have been processed. The current format is compatible with Graphite.

Systemd integration

haproxystats-pull and haproxystats-process are simple programs which are not daemonized and they output logging messages to stdout. This is by design as it simplifies the code. The daemonization and logging is off-loaded to systemd which has everything we need for that job.

Under contrib/systend directory there are service files for both programs. These are functional systemd Unit files which are used in production.

The order in which these 2 programs start doesn’t matter and there isn’t any soft or hard dependency between them.

Furthermore, these programs don’t need to run as root. It highly recommended to create a dedicated user to run them. You need to add that user to the group of haproxy and adjust socket configuration of haproxy to allow write for the group, see below an example configuration:

stats socket /run/haproxy/sock1 user haproxy group haproxy mode 660 level admin process 1
stats socket /run/haproxy/sock2 user haproxy group haproxy mode 660 level admin process 2
stats socket /run/haproxy/sock3 user haproxy group haproxy mode 660 level admin process 3

systemd Unit files use haproxystats user which has to be created prior running haproxystats programs.

Graceful shutdown

In an effort to reduce the loss of statistics both programs support graceful shutdown. When SIGHUP or SIGTERM signals are sent they perform a clean exit. When a signal is sent to haproxystats-process it may take some time for the program to exit, as it waits for all workers to empty the queue.

Puppet module

A puppet module is available under contrib directory which provides classes for configuring both programs.

Because haproxystats-process is CPU bound program, CPU Affinity is configured using systemd. By default it pins the workers to the last CPUs.

You should take care of pinning haproxy processes to other CPUs in order to avoid haproxystats-process stealing CPU cycles from haproxy. In production servers you usually pin the first 80% of CPUs to haproxy processes and you leave the rest of CPUs for other processes. The default template of puppet module enforces this logic.

haproxystats-pull is a single threaded program which doesn’t use a lot of CPU cycles and by default is assigned to the last CPU.

Ansible Playbook

A Ansible playbook is available under contrib directory. For installation instruction of the playbook please read Installation chapter of this document.

Nagios checks

Several nagios checks are provided for monitoring purposes, they can be found under contrib/nagios directory.

  • check_haproxystats_process_number_of_procs.sh

Monitor the number of processes of haproxystats-process program. Systemd monitors only the parent process and this check helps to detect cases where some worker(s) die unexpectedly

  • check_haproxystats_process.sh

A wrapper around systemctl tool to detect a dead parent process.

  • check_haproxystats_pull.sh

A wrapper around systemctl tool to a check if haproxystats-pull is running.

  • check_haproxystats_queue_size.py

Checks the size of the incoming directory queue which is consumed by haproxystats-process and alert when exceeds a threshold.

Monit check

If a child process of haproxystats-process dies then monit can restart haproxystats-process. There is a monit check configuration available under contrib/monit directory which does that.

Starting the programs

haproxystats-pull -f ./haproxystats.conf
haproxystats-process -f ./haproxystats.conf

Usage:

% haproxystats-pull -h
Pulls statistics from HAProxy daemon over UNIX socket(s)

Usage:
    haproxystats-pull [-f <file> ] [-p | -P]

Options:
    -f, --file <file>  configuration file with settings
                       [default: /etc/haproxystats.conf]
    -p, --print        show default settings
    -P, --print-conf   show configuration
    -h, --help         show this screen
    -v, --version      show version


% haproxystats-process -h
Processes statistics from HAProxy and pushes them to Graphite

Usage:
    haproxystats-process [-f <file> ] [-p | -P]

Options:
    -f, --file <file>  configuration file with settings
                       [default: /etc/haproxystats.conf]
    -p, --print        show default settings
    -P, --print-conf   show configuration
    -h, --help         show this screen
    -v, --version      show version

Development

I would love to hear what other people think about haproxystats and provide feedback. Please post your comments, bug reports and wishes on my issues page.

How to setup a development environment

Install HAProxy:

% sudo apt-get install haproxy

Use a basic HAProxy configuration in multiprocess mode:

global
    log 127.0.0.1 len 2048 local2
    chroot /var/lib/haproxy
    stats socket /run/haproxy/admin1.sock mode 666 level admin process 1
    stats socket /run/haproxy/admin2.sock mode 666 level admin process 2
    stats socket /run/haproxy/admin3.sock mode 666 level admin process 3
    stats socket /run/haproxy/admin4.sock mode 666 level admin process 4
    # allow read/write access to anyone----------^
    stats timeout 30s
    user haproxy
    group haproxy
    daemon
    nbproc 4
    cpu-map 1 0
    cpu-map 2 1
    cpu-map 3 1
    cpu-map 4 0

defaults
    log global
    mode    http
    timeout connect 5000
    timeout client  50000
    timeout server  50000

frontend frontend_proc1
    bind 0.0.0.0:81 process 1
    default_backend backend_proc1

frontend frontend_proc2
    bind 0.0.0.0:82 process 2
    default_backend backend_proc1

frontend frontend1_proc34
    bind :83 process 3
    bind :83 process 4
    default_backend backend1_proc34

backend backend_proc1
    bind-process 1
    default-server inter 1000s
    option httpchk GET / HTTP/1.1\r\nHost:\ .com\r\nUser-Agent:\ HAProxy
    server member1_proc1 10.189.224.169:80 weight 100 check fall 2 rise 3
    server member2_proc1 10.196.70.109:80 weight 100 check fall 2 rise 3
    server bck_all_srv1 10.196.70.109:88 weight 100 check fall 2 rise 3

backend backend1_proc34
    bind-process 3,4
    default-server inter 1000s
    option httpchk GET / HTTP/1.1\r\nHost:\ .com\r\nUser-Agent:\ HAProxy
    server bck1_proc34_srv1 10.196.70.109:80 check fall 2 inter 5s rise 3
    server bck1_proc34_srv2 10.196.70.109:80 check fall 2 inter 5s rise 3
    server bck_all_srv1 10.196.70.109:80 check fall 2 inter 5s rise 3

backend backend_proc2
    bind-process 2
    default-server inter 1000s
    option httpchk GET / HTTP/1.1\r\nHost:\ .com\r\nUser-Agent:\ HAProxy
    server bck_proc2_srv1_proc2 127.0.0.1:8001 check fall 2 inter 5s rise 3
    server bck_proc2_srv2_proc2 127.0.0.1:8002 check fall 2 inter 5s rise 3
    server bck_proc2_srv3_proc2 127.0.0.1:8003 check fall 2 inter 5s rise 3
    server bck_proc2_srv4_proc2 127.0.0.1:8004 check fall 2 inter 5s rise 3

Start HAProxy and check it is up:

sudo systemctl start haproxy.service;systemctl status -l haproxy.service

Create a python virtual environment using virtualenvwrapper tool:

mkvirtualenv --python=`which python3` haproxystats-dev

Do not exit the haproxystats-dev virtual environment.

Clone the project, if you are planning to contribute then you should fork it on GitHub and clone that project instead:

mkdir ~/repo;cd ~/repo
git clone https://github.com/unixsurfer/haproxystats

Install necessary libraries:

cd haproxystats
pip install -U pbr setuptools
pip install -r ./requirements.txt

Start a TCP server which acts a Graphite relay and listens on 127.0.0.1:39991:

python3 ./contrib/tcp_server.py

Install haproxystats:

python setup.py install

Create necessary directory structure:

mkdir -p ./var/var/lib/haproxystats
mkdir -p ./var/etc
mkdir -p ./var/etc/haproxystats.d

Adjust the following configuration and save it in ./var/etc/haproxystats.conf:

[DEFAULT]
loglevel = debug
retries  = 2
timeout  = 1
interval = 2

[paths]
base-dir = /home/<username>/repo/haproxystats/var/var/lib/haproxystats

[pull]
socket-dir    = /run/haproxy
retries       = 1
timeout       = 0.1
interval      = 0.5
pull-timeout  = 10
pull-interval = 10
dst-dir       = ${paths:base-dir}/incoming
tmp-dst-dir   = ${paths:base-dir}/incoming.tmp
workers       = 8

[process]
src-dir               = ${paths:base-dir}/incoming
workers               = 2
calculate-percentages = true
per-process-metrics   = true

[graphite]
server          = 127.0.0.1
port            = 39991
retries         = 3
interval        = 0.8
timeout         = 0.9
delay           = 10
backoff         = 2
namespace       = loadbalancers
prefix_hostname = true
fqdn            = true
queue-size      = 1000

[local-store]
dir = ${paths:base-dir}/local-store

Start haproxystats-pull and haproxystats-process on 2 different terminals:

haproxystats-pull -f var/etc/haproxystats.conf
haproxystats-process -f var/etc/haproxystats.conf

Exit from haproxystats-dev virtual environment:

deactivate

Start hacking and don’t forget to make a Pull Request

Installation

Use pip:

pip install haproxystats

From Source:

sudo python setup.py install

Build (source) RPMs:

python setup.py clean --all; python setup.py bdist_rpm

Build a source archive for manual installation:

python setup.py sdist

Use Ansible Playbook:

To deploy haproxystats By Ansible Playbook go to contrib/ansible-playbook directory:

cd contrib/ansible-playbook

Then enter your haproxy server IP address in hosts file:

vi hosts

After that set information of your environment in variable file:

vi group_vars/all

Now for run Ansible Playbook use this command:

ansible-playbook -i hosts main-playbook.yml

When Ansible Playbook run successful completely, you can take control haproxystats-pull and haproxystats-process by systemd:

systemctl start haproxystats-pull.service

systemctl start haproxystats-process.service

How to make a release

  1. Bump version in haproxystats/__init__.py

  2. Commit above change with:

    git commit -av -m'RELEASE 0.1.3 version'
  3. Create a signed tag, pbr will use this for the version number:

    git tag -s 0.1.3 -m 'bump release'
  4. Create the source distribution archive (the archive will be placed in the dist directory):

    python setup.py sdist
  5. pbr updates ChangeLog file and we want to squeeze this change to the previous commit, thus run:

    git commit -av --amend
  6. Move current tag to the last commit:

    git tag -fs 0.1.3 -m 'bump release'
  7. Push changes:

    git push;git push --tags
  8. Upload to Python Package Index:

    twine upload -s -p  dist/*

Contributors

The following people have contributed to project with feedback and code reviews

Licensing

Apache 2.0

Acknowledgement

This program was originally developed for Booking.com. With approval from Booking.com, the code was generalised and published as Open Source on github, for which the author would like to express his gratitude.

Contacts

Project website: https://github.com/unixsurfer/haproxystats

Author: Pavlos Parissis <pavlos.parissis@gmail.com>

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

haproxystats-0.4.1.tar.gz (127.3 kB view hashes)

Uploaded Source

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page