playing with metrics
Project description
pytheus
playing with metrics
Documentation: https://pythe.us
pytheus is a modern python library for collecting prometheus metrics built with multiprocessing in mind.
Some of the features are:
- multiple multiprocess support:
- redis backend ✅
- Rust powered backend 🧪
- bring your own ✅
- support for default labels value ✅
- partial labels value (built in an incremental way) ✅
- customizable registry support ✅
- registry prefix support ✅
Philosophy
Simply put is to let you work with metrics the way you want.
Be extremely flexible, allow for customization from the user for anything they might want to do without having to resort to hacks and most importantly offer the same api for single & multi process scenarios, the switch should be as easy as loading a different backend without having to change anything in the code.
- What you see is what you get.
- No differences between
singleprocess
&multiprocess
, the only change is loading a different backend and everything will work out of the box. - High flexibility with an high degree of
labels
control and customization.
Requirements
- Python 3.8+
- redis >= 4.0.0 (optional: for multiprocessing)
- pytheus-backend-rs (optional: for Rust powered multiprocessing 🦀)
Architecture
A small piece of the architecture is described in the ARCHITECTURE.md file
Install
pip install pytheus
Optionally if you want to use the Redis backend (for multiprocess support) you will need the redis library:
pip install redis
# or
pip install pytheus[redis]
If you want to try the Rust based backend (for multiprocess support):
pip install pytheus-backend-rs
Partial labels support:
from pytheus.metrics import Counter
# without labels
my_metric = Counter('metric_name', 'desc')
my_metric.inc() # example for counter
# with labels
my_metric = Counter('metric_name', 'desc', required_labels=['req1', 'req2'])
my_metric.labels({'req1': '1', 'req2': '2'}).inc() # you can pass all the labels at once
partial_my_metric = my_metric.labels({'req1': '1'}) # a cacheable object with one of the required labels already set
observable_my_metric = partial_my_metric.labels({'req2': '2'}) # finish setting the remaining values before observing
observable_my_metric.inc()
# alternatively to using a `dict` you can use `kwargs` (keyword arguments)
my_metric.labels(req1='1', req2='2')
Default labels support:
from pytheus.metrics import Counter
# with default labels
my_metric = Counter('metric_name', 'desc', required_labels=['req1', 'req2'], default_labels={'req2': 'me_set!'})
my_metric.labels({'req1': '1'}).inc() # as we have req2 as a default label we only need to set the remaining labels for observing
my_metric.labels({'req1': '1', 'req2': '2'}) # you can still override default labels!
Exposing metrics:
You can use the generate_metrics
function from pytheus.exposition
to generate the metrics and serve them as an endpoint with your favourite web framework.
Alternatively you can use the make_wsgi_app
function that creates a simple wsgi app to serve the metrics.
Quickstart / Example
The example.py
file starts a flask application with three endpoints:
/
: just returns a phrase while observing the time taken for the request to complete/slow
: same as before but will sleep so that values will only end up in higher buckets/metrics
: the endpoint used by prometheus to scrape the metrics
It uses two histograms, one without labels, and one with labels required and a default label that makes it observable.
To expose the metrics the generate_metrics()
function is used.
note: the example file is using the redis backend but you can try without and set up prometheus yourself.
Redis version
For the redis version you can just clone the repository and run docker-compose up
to start both redis and prometheus scraping on localhost:8080.
Then you can start the local server with python example.py
. (flask is required for it to work)
Now you can visit the described endpoints and by visiting localhost:9090
you can query prometheus, for example by looking for all the slow requests buckets: page_visits_latency_seconds_labeled_bucket{speed="slow"}
Default version
For the default single process version you can create your python server like this:
import time
from flask import Flask
from pytheus.metrics import Histogram
from pytheus.exposition import generate_metrics
app = Flask(__name__)
histogram = Histogram('page_visits_latency_seconds', 'used for testing')
# this is the endpoint that prometheus will use to scrape the metrics
@app.route('/metrics')
def metrics():
return generate_metrics()
# track time with the context manager
@app.route('/')
def home():
with histogram.time():
return 'hello world!'
# you can also track time with the decorator shortcut
@app.route('/slow')
@histogram
def slow():
time.sleep(3)
return 'hello world! from slow!'
app.run(host='0.0.0.0', port=8080)
and if you have prometheus installed configure it to scrape on localhost:8080 or you can still use the included docker-compose.yml
file.
Metric types
Counter
The Counter is a metric that only increases and can resets to 0. (For example if a service restart, it will start again from zero)
from pytheus.metrics import Counter
counter = Counter(name="my_counter", description="My description")
# increase by 1
counter.inc()
# increase by x
counter.inc(7)
# it is possible to count exceptions
with counter.count_exceptions():
raise ValueError # increases counter by 1
# you can specify which exceptions to watch for
with counter.count_exceptions((IndexError, ValueError)):
raise ValueError. # increases counter by 1
# it is possible to use the counter as a decorator as a shortcut to count exceptions
@counter
def test():
raise ValueError # increases counter by 1 when called
# specifying which exceptions to look for also works with the decorator
@counter(exceptions=(IndexError, ValueError))
def test():
raise ValueError # increases counter by 1 when called
Gauge
The Gauge can increase and decrease its value. It is also possible to set a specific value.
from pytheus.metrics import Gauge
gauge = Gauge(name="my_gauge", description="My description")
# increase by 1
gauge.inc()
# increase by x
gauge.inc(7)
# decrease by 1
gauge.dec()
# set a specific value
gauge.set(7)
# set to current unix timestamp
gauge.set_to_current_time()
# it is possible to track progress so that when entered increases the value by 1, and when exited decreases it
with gauge.track_inprogress():
do_something()
# you can also time a piece of code, will set the duration in seconds to value when exited
with gauge.time():
do_something()
# tracking time can also be done as a decorator
@gauge
def do_something():
...
# tracking progress is also available via decorator with a flag
@gauge(track_inprogress=True)
def do_something():
...
Histogram
A histogram samples observations (usually things like request durations or response sizes) and counts them in configurable buckets. It also provides a sum of all observed values. (taken from prometheus docs)
from pytheus.metrics import Histogram
histogram = Histogram(name="my_histogram", description="My description")
# by default it will have the following buckets: (.005, .01, .025, .05, .1, .25, .5, 1, 2.5, 5, 10)
# note: the +Inf bucket will be added automatically, this is float('inf') in python
# create a histogram specifying buckets
histogram = Histogram(name="my_histogram", description="My description", buckets=(0.2, 1, 3))
# observe a value
histogram.observe(0.4)
# you can also time a piece of code, will set the duration in seconds to value when exited
with histogram.time():
do_something()
# tracking time can also be done as a decorator
@histogram
def do_something():
...
Summary
A Summary is a metric that captures individual observations and tracks the total size & number of events observed. It can be useful to track latencies for example.
from pytheus.metrics import Summary
summary = Summary(name="my_summary", description="My description")
# observe a value
summary.observe(0.4)
# you can also time a piece of code, will set the duration in seconds to value when exited
with summary.time():
do_something()
# tracking time can also be done as a decorator
@summary
def do_something():
...
Custom Collectors
It is possible to use a custom collector in cases you can't directly instrument the code.
You will need to inherit from CustomCollector
and define the collect
method.
Also be sure to disable the automatic registering of a newly created metric into the default registry.
from pytheus.metrics import Counter, CustomCollector
from pytheus.registry import REGISTRY
class MyCustomCollector(CustomCollector):
def collect():
counter = Counter('name', 'desc', registry=None) # note that we are disabling automatic registering
counter.inc()
yield counter
REGISTRY.register(MyCustomCollector())
Note: if one of the yield metrics have a name already registered with the registry you are trying to register to, the custom collector will be ignored.
How to use different backends
Things work out of the box, using the SingleProcessBackend:
from pytheus.metrics import Counter
counter = Counter(
name="my_metric",
description="My description",
required_labels=["label_a", "label_b"],
)
print(counter._metric_value_backend.__class__)
# <class 'pytheus.backends.base.SingleProcessBackend'>
print(counter._metric_value_backend.config)
# {}
You can define environment configuration to have different defaults, using two environment variables:
export PYTHEUS_BACKEND_CLASS="pytheus.backends.redis.MultiProcessRedisBackend"
export PYTHEUS_BACKEND_CONFIG="./config.json"
Now, create the config file, ./config.json
:
{
"host": "127.0.0.1",
"port": 6379
}
Now we can try the same snippet as above:
from pytheus.metrics import Counter
counter = Counter(
name="my_metric",
description="My description",
required_labels=["label_a", "label_b"],
)
print(counter._metric_value_backend.__class__)
# <class 'pytheus.backends.redis.MultiProcessRedisBackend'>
print(counter._metric_value_backend.config)
# {"host": "127.0.0.1", "port": 6379}
You can also pass the values directly in Python, which would take precedence over the environment setup we have just described:
from pytheus.metrics import Counter
from pytheus.backends import load_backend
from pytheus.backends.redis import MultiProcessRedisBackend
load_backend(
backend_class=MultiProcessRedisBackend,
backend_config={
"host": "127.0.0.1",
"port": 6379
}
)
# Notice that if you simply call load_backend(), it would reload config from the environment.
# load_backend() is called automatically at package import, that's why we didn't need to call it
# directly in the previous example
counter = Counter(
name="my_metric",
description="My description",
required_labels=["label_a", "label_b"],
)
print(counter._metric_value_backend.__class__)
# <class 'pytheus.backends.redis.MultiProcessRedisBackend'>
print(counter._metric_value_backend.config)
# {'host': '127.0.0.1', 'port': 6379}
Create your own Backend
Custom Backend
You can create your own backend by implementing a class that fulfills the Backend
protocol.
class Backend(Protocol):
def __init__(
self,
config: BackendConfig,
metric: "_Metric",
histogram_bucket: str | None = None,
) -> None:
...
def inc(self, value: float) -> None:
...
def dec(self, value: float) -> None:
...
def set(self, value: float) -> None:
...
def get(self) -> float:
...
Initialization hook
It's possible that you want to initialize your custom backend or there are one time steps that you want to happen on import.
To achieve that you can use the class method hook called _initialize
that accepts a BackendConfig
parameter.
@classmethod
def _initialize(cls, config: "BackendConfig") -> None:
# initialization steps
Client patching
An highly experimental feature is to patch in-place the official prometheus_client
so that you can test out the multiprocessing mode of this library without committing to it or just to try it out.
import prometheus_client
from pytheus.experimental.compatibility import patch_client
patch_client(prometheus_client)
This has to be done at the start/initialization of your application and you can load a different backend before applying with patch with patch_client
by using load_backend
.
This supports all the metrics types Counter
, Gauge
, Summary
& Histogram
. It supports the default REGISTRY
and generate_latest
. It also works with the start_http_server
function.
Of course it's extremely limited and you might be using a feature that is not yet supported, but feel free to try! :)
(This was used to test an internal library without having to migrate all of it, so since I worked on it might as well share it! )
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file pytheus-0.6.0.tar.gz
.
File metadata
- Download URL: pytheus-0.6.0.tar.gz
- Upload date:
- Size: 34.6 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.1 CPython/3.10.14
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 0d0dd7d7495de8cce2e1c63c15b75f3daa22da52f43ffd1370a40e91ed368f1d |
|
MD5 | ecc1a3d7dc431707352eb6954fe0660e |
|
BLAKE2b-256 | fc414f63ed7d8702d3e2b77d8ec545aae20487a3bbbc954207ebc500f42ccf22 |
File details
Details for the file pytheus-0.6.0-py3-none-any.whl
.
File metadata
- Download URL: pytheus-0.6.0-py3-none-any.whl
- Upload date:
- Size: 26.8 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.1 CPython/3.10.14
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 077e859b1a057066c9a91a21e6296c242a18f8efbe693212843d2207ed1981d8 |
|
MD5 | f5f41e03953c04b4be7a7a62a2c206f4 |
|
BLAKE2b-256 | eba9b043f7ac4267b294c310dea822b2ddce2ae82c39825297b14ef3074e1571 |