Various python stuff: testing, aio helpers, etc
Project description
# About this package
[![Build Status](https://travis-ci.org/tipsi/tipsi_tools.svg?branch=master)](https://travis-ci.org/tipsi/tipsi_tools)
[![PyPi version](https://img.shields.io/pypi/v/tipsi_tools.svg)](https://pypi.python.org/pypi/tipsi_tools)
Here are set of internal tools that are shared between different projects internally. Originally most tools related to testing, so they provide some base classes for various cases in testing
**NOTE: all our tools are intentially support only 3.5+ python.**
Some might work with other versions, but we're going to be free from all these crutches to backport things like `async/await` to lower versions, so if it works - fine, if not - feel free to send PR, but it isn't going to be merged all times.
## Testing helpers
### ApiUrls
Defined in `tipsi_tools/testing/__init__.py`. Required for defining nested urls with formatting.
You can use it in fixtures, like:
```python
@pytest.fixture(scope='session')
def api(api_v_base):
yield ApiUrls('{}/'.format(api_v_base), {
'password_reset_request': 'password/request/code/',
'password_reset': 'password/reset/',
'user_review_list': 'user/{user_id}/review/',
'user_review': 'user/{user_id}/review/{review_id}/',
'wine_review': 'wine/{wine_id}/review/',
'drink_review': 'drink/{drink_id}/review/',
})
def test_review_list(user, api):
resp = user.get_json(api.user_review_list(user_id=user1.id), {'page_size': 2})
```
### PropsMeta
You can find source in `tipsi_tools/testing/meta.py`.
For now it convert methods that are started with `prop__` into descriptors with cache.
```python
class A(metaclass=PropsMeta):
def prop__conn(self):
conn = SomeConnection()
return conn
```
Became:
```python
class A:
@property
def conn(self):
if not hasattr(self, '__conn'):
setattr(self, '__conn', SomeConnection())
return self.__conn
```
Thus it allows quite nice style of testing with lazy initialization. Like:
```python
class MyTest(TestCase, metaclass=PropsMeta):
def prop__conn(self):
return psycopg2.connect('')
def prop__cursor(self):
return self.conn.cursor()
def test_simple_query(self):
self.cursor.execute('select 1;')
row = self.cursor.fetchone()
assert row[0] == 1, 'Row: {}'.format(row)
```
Here you just get and use `self.cursor`, but automatically you get connection and cursor and cache they.
This is just simple example, complex tests can use more deep relations in tests. And this approach is way more easier and faster than complex `setUp` methods.
### AIOTestCase
**NOTE: we're highly suggest to use pytest alongside with existing async tests plugins**
Base for asyncronous test cases, you can use it as drop-in replacement for pre-existent tests to be able:
* write asyncronous test methods
* write asyncronous `setUp` and `tearDown` methods
* use asyncronous function in `assertRaises`
```python
class ExampleCase(AIOTestCase):
async setUp(self):
await async_setup()
async tearDown(self):
await async_teardown()
async division(self):
1/0
async test_example(self):
await self.assertRaises(ZeroDivisionError, self.async_division)
```
## tipsi_tools.unix helpers
Basic unix helpers
* run - run command in shell
* succ - wrapper around `run` with return code and stderr check
* wait_socket - wait for socket awailable (eg. you can wait for postgresql with `wait_socket('localhost', 5432)`
* asucc - asynchronous version of `succ` for use with `await`. supports realtime logging
* source - acts similar to bash 'source' or '.' commands.
* cd - contextmanager to do something with temporarily changed directory
#### interpolate_sysenv
Format string with system variables + defaults.
```python
PG_DEFAULTS = {
'PGDATABASE': 'postgres',
'PGPORT': 5432,
'PGHOST': 'localhost',
'PGUSER': 'postgres',
'PGPASSWORD': '',
}
DSN = interpolate_sysenv('postgresql://{PGUSER}:{PGPASSWORD}@{PGHOST}:{PGPORT}/{PGDATABASE}', PG_DEFAULTS)
```
## tipsi_tools.logging.JSFormatter
Enable json output with additional fields, suitable for structured logging into ELK or similar solutions.
Accepts `env_vars` key with environmental keys that should be included into log.
```python
# this example uses safe_logger as handler (pip install safe_logger)
import logging
import logging.config
LOGGING = {
'version': 1,
'disable_existing_loggers': True,
'formatters': {
'json': {
'()': 'tipsi_tools.logging.JSFormatter',
'env_vars': ['HOME'],
},
'standard': {
'format': '%(asctime)s [%(levelname)s] %(name)s: %(message)s'
},
},
'handlers': {
'default': {
'level': 'DEBUG',
'class': 'safe_logger.TimedRotatingFileHandlerSafe',
'filename': 'test_json.log',
'when': 'midnight',
'interval': 1,
'backupCount': 30,
'formatter': 'json',
},
},
'loggers': {
'': {
'handlers': ['default'],
'level': 'DEBUG',
},
},
}
logging.config.dictConfig(LOGGING)
log = logging.getLogger('TestLogger')
log.debug('test debug')
log.info('test info')
log.warn('test warn')
log.error('test error')
```
## tipsi_tools.mon_server.MetricsServer
Sanic based server that servers metrics in prometheus format.
```
from sanic import Sanic
from tipsi_tools.mon_server import MetricsServer
from tipsi_tools.mon_server.certs import update_certs_loop
app = Sanic()
mserver = MetricsServer(app, status_metric='running{server="localhost"}')
mserver.add_task(update_certs_loop, hosts=['gettipsi.com', 'proofnetwork.io'])
app.run(host='0.0.0.0', port=8000)
```
## tipsi_tools.drf.serializers.EnumSerializer
Allow you to deserealize incoming strings into `Enum` values.
You should add `EnumSerializer` into your serializers by hand.
```python
from enum import IntEnum
from django.db import models
from rest_framework import serializers
from tipsi_tools.drf.serializers import EnumSerializer
class MyEnum(IntEnum):
one = 1
two = 2
class ExampleModel(models.Model):
value = models.IntegerField(choices=[(x.name, x.value) for x in MyEnum])
class ExampleSerializer(serializers.ModelSerializer):
value = EnumSerializer(MyEnum)
# this allows you to post value as: {'value': 'one'}
```
Due to `Enum` and `IntegerField` realizations you may use `Enum.value` in querysets
```python
ExampleModel.objects.filter(value=MyEnum.two)
```
## tipsi_tools.django.log_requests.LoggerMiddleware
LoggerMiddleware will log request meta + raw post data into log.
For django<1.10 please use `tipsi_tools.django.log_requests.DeprecatedLoggerMiddleware`
## tipsi_tools.django.request_uniq
Decorator adds a unique for each uwsgi request dict as first function
argument.
For tests mock `_get_request_unique_cache`
## tipsi_tools.django.call_once_on_commit
Make function called only once on transaction commit. Here is examples
where function `do_some_useful` will be called only once after
transaction has been committed.
```python
class SomeModel(models.Model):
name = IntegerField()
@call_once_on_commit
def do_some_useful():
pass
def hook(sender, instance, **kwargs):
do_some_useful()
models.signals.post_save.connect(hook, sender=SomeModel)
with transaction.atomic():
some_model = SomeModel()
some_model.name = 'One'
some_model.save()
some_model.name = 'Two'
some_model.save()
```
For tests with nested transactions (commit actually most times is not
called) it is useful to override behaviour `call_once_on_commit`
when decorated function executed right in place where it is called.
To do so mock `on_commit` function. Example pytest fixture:
```
@pytest.fixture(scope='session', autouse=True)
def immediate_on_commit():
def side_effect():
return lambda f: f()
with mock.patch('tipsi_tools.django.on_commit', side_effect=side_effect) as m:
yield m
```
## tipsi_tools.django.fields.ChoicesEnum
Used for choices attribute for in model field
```
class FooBarEnum(ChoicesEnum):
foo = 1
bar = 2
class ExampleModel(models.Model):
type = models.IntegerField(choices=FooBarEnum.get_choices())
```
## tipsi_tools.django.db.utils.set_word_similarity_threshold
Allow to set postgres trigram word similarity threshold for default django database connection
```
set_word_similarity_threshold(0.4)
```
## tipsi_tools.django.db.pgfields.LTreeField
Django postgres ltree field type
```
class LTreeExample(models.Model):
path = LTreeField()
```
## tipsi_tools.django.db.pgfields.LTreeDescendant
Lookup for postgres ltree
```
# Add this import to models.py (file should be imported before lookup usage)
import tipsi_tools.django.db.pgfields # noqa
LTreeExample.objects.filter(path__descendant='root.level1')
```
## tipsi_tools.django.db.pgfields.SimilarityLookup
Postgres `text %> text` operator
```
# Add this import to models.py (file should be imported before lookup usage)
import tipsi_tools.django.db.pgfields # noqa
Books.objects.filter(title__similar='Animal Farm')
```
## tipsi_tools.django.db.pgfields.WordSimilarity
Postgres `text1 <<-> text2` operator. It returns `1 - word_similarity(text1, text2)`
```
from django.db.models import Value, F
similarity = WordSimilarity(Value('Animal Farm'), F('title'))
Books.objects.annotate(similarity=similarity)
```
## tipsi_tools.drf.use_form
Helps to use power of serializers for simple APIs checks.
```python
from rest_framework import serializers
from rest_framework.decorators import api_view
from tipsi_tools.drf import use_form
class SimpleForm(serializers.Serializer):
test_int = serializers.IntegerField()
test_str = serializers.CharField()
@api_view(['GET'])
@use_form(SimpleForm)
def my_api(data):
print(f'Data: {data["test_int"]} and {data["test_str"]}')
```
## tipsi_tools.aio_utils.DbRecordsProcessorWorker
Asyncio worker which wait for new records in postgres db table and process them.
## tipsi_tools.aio_utils.dict_query/sql_update
aiopg shortcuts
## tipsi_tools.python.execfile
Backport of python's 2 `execfile` function.
Usage: execfile('path/to/file.py', globals(), locals())
Returns: True if file exists and executed, False if file doesn't exist
## tipsi_tools.doc_utils.tipsi_sphinx
Sphinx extensions to generate documentation for django restframework serializers and examples for http requests.
In order to use them specify dependency for package installation:
```
pip install tipsi_tools[doc_utils]
```
Usage:
```
# Add to Sphinx conf.py
extensions = [
# ...
'tipsi_tools.doc_utils.tipsi_sphinx.dyn_serializer',
'tipsi_tools.doc_utils.tipsi_sphinx.http_log'
]
```
## Commands
### tipsi_env_yaml
Convert template yaml with substituion of `%{ENV_NAME}` strings to appropriate environment variables.
Usage: `tipsi_env_yaml src_file dst_file`
### tipsi_ci_script
Helper to run default CI pipeline. Defaults are set up for [giltab defaults](https://docs.gitlab.com/ee/ci/variables/#predefined-variables-environment-variables). Includes stages:
* build docker image with temporary name (commit sha by default)
* run tests (optional)
* push branch (by default only for master and staging branches)
* push tag, if there are tags
* cache image with common name
* delete image with temporary name
It's optimized for parallel launches, so you need to use unique temporary name (`--temp-name`). We want keep our system clean if possible, so we'll delete this tag in the end. But we don't want to repeat basic steps over and over, so we will cache image with common cache name (`--cache-name`), it will remove previous cached image.
### tipsi_wait
Wait for socket awailable/not-awailable with timeout.
```
# Wait until database port up for 180 seconds
tipsi_wait -t 180 postgres 5432
# Wait until nginx port down for 30 seconds
tipsi_wait -t 30 nginx 80
```
### run_filebeat
* checks environmental variables `-e KEY=VALUE -e KEY2=VALUE2`
* converts yaml template `tipsi_env_yaml {TEMPLATE} /tmp/filebeat.yml`
* run `/usr/bin/filebeat /tmp/filebeat.yml`
```
run_filebeat -e CHECKME=VALUE path_to_template
```
### doc_serializer
* output rst with list of serializers
* generates documentation artifacts for serializers
```
usage: doc_serializer [-h] [--rst] [--artifacts]
Parse serializers sources
optional arguments:
-h, --help show this help message and exit
--rst Output rst with serializers
--artifacts Write serializers artifacts
```
[![Build Status](https://travis-ci.org/tipsi/tipsi_tools.svg?branch=master)](https://travis-ci.org/tipsi/tipsi_tools)
[![PyPi version](https://img.shields.io/pypi/v/tipsi_tools.svg)](https://pypi.python.org/pypi/tipsi_tools)
Here are set of internal tools that are shared between different projects internally. Originally most tools related to testing, so they provide some base classes for various cases in testing
**NOTE: all our tools are intentially support only 3.5+ python.**
Some might work with other versions, but we're going to be free from all these crutches to backport things like `async/await` to lower versions, so if it works - fine, if not - feel free to send PR, but it isn't going to be merged all times.
## Testing helpers
### ApiUrls
Defined in `tipsi_tools/testing/__init__.py`. Required for defining nested urls with formatting.
You can use it in fixtures, like:
```python
@pytest.fixture(scope='session')
def api(api_v_base):
yield ApiUrls('{}/'.format(api_v_base), {
'password_reset_request': 'password/request/code/',
'password_reset': 'password/reset/',
'user_review_list': 'user/{user_id}/review/',
'user_review': 'user/{user_id}/review/{review_id}/',
'wine_review': 'wine/{wine_id}/review/',
'drink_review': 'drink/{drink_id}/review/',
})
def test_review_list(user, api):
resp = user.get_json(api.user_review_list(user_id=user1.id), {'page_size': 2})
```
### PropsMeta
You can find source in `tipsi_tools/testing/meta.py`.
For now it convert methods that are started with `prop__` into descriptors with cache.
```python
class A(metaclass=PropsMeta):
def prop__conn(self):
conn = SomeConnection()
return conn
```
Became:
```python
class A:
@property
def conn(self):
if not hasattr(self, '__conn'):
setattr(self, '__conn', SomeConnection())
return self.__conn
```
Thus it allows quite nice style of testing with lazy initialization. Like:
```python
class MyTest(TestCase, metaclass=PropsMeta):
def prop__conn(self):
return psycopg2.connect('')
def prop__cursor(self):
return self.conn.cursor()
def test_simple_query(self):
self.cursor.execute('select 1;')
row = self.cursor.fetchone()
assert row[0] == 1, 'Row: {}'.format(row)
```
Here you just get and use `self.cursor`, but automatically you get connection and cursor and cache they.
This is just simple example, complex tests can use more deep relations in tests. And this approach is way more easier and faster than complex `setUp` methods.
### AIOTestCase
**NOTE: we're highly suggest to use pytest alongside with existing async tests plugins**
Base for asyncronous test cases, you can use it as drop-in replacement for pre-existent tests to be able:
* write asyncronous test methods
* write asyncronous `setUp` and `tearDown` methods
* use asyncronous function in `assertRaises`
```python
class ExampleCase(AIOTestCase):
async setUp(self):
await async_setup()
async tearDown(self):
await async_teardown()
async division(self):
1/0
async test_example(self):
await self.assertRaises(ZeroDivisionError, self.async_division)
```
## tipsi_tools.unix helpers
Basic unix helpers
* run - run command in shell
* succ - wrapper around `run` with return code and stderr check
* wait_socket - wait for socket awailable (eg. you can wait for postgresql with `wait_socket('localhost', 5432)`
* asucc - asynchronous version of `succ` for use with `await`. supports realtime logging
* source - acts similar to bash 'source' or '.' commands.
* cd - contextmanager to do something with temporarily changed directory
#### interpolate_sysenv
Format string with system variables + defaults.
```python
PG_DEFAULTS = {
'PGDATABASE': 'postgres',
'PGPORT': 5432,
'PGHOST': 'localhost',
'PGUSER': 'postgres',
'PGPASSWORD': '',
}
DSN = interpolate_sysenv('postgresql://{PGUSER}:{PGPASSWORD}@{PGHOST}:{PGPORT}/{PGDATABASE}', PG_DEFAULTS)
```
## tipsi_tools.logging.JSFormatter
Enable json output with additional fields, suitable for structured logging into ELK or similar solutions.
Accepts `env_vars` key with environmental keys that should be included into log.
```python
# this example uses safe_logger as handler (pip install safe_logger)
import logging
import logging.config
LOGGING = {
'version': 1,
'disable_existing_loggers': True,
'formatters': {
'json': {
'()': 'tipsi_tools.logging.JSFormatter',
'env_vars': ['HOME'],
},
'standard': {
'format': '%(asctime)s [%(levelname)s] %(name)s: %(message)s'
},
},
'handlers': {
'default': {
'level': 'DEBUG',
'class': 'safe_logger.TimedRotatingFileHandlerSafe',
'filename': 'test_json.log',
'when': 'midnight',
'interval': 1,
'backupCount': 30,
'formatter': 'json',
},
},
'loggers': {
'': {
'handlers': ['default'],
'level': 'DEBUG',
},
},
}
logging.config.dictConfig(LOGGING)
log = logging.getLogger('TestLogger')
log.debug('test debug')
log.info('test info')
log.warn('test warn')
log.error('test error')
```
## tipsi_tools.mon_server.MetricsServer
Sanic based server that servers metrics in prometheus format.
```
from sanic import Sanic
from tipsi_tools.mon_server import MetricsServer
from tipsi_tools.mon_server.certs import update_certs_loop
app = Sanic()
mserver = MetricsServer(app, status_metric='running{server="localhost"}')
mserver.add_task(update_certs_loop, hosts=['gettipsi.com', 'proofnetwork.io'])
app.run(host='0.0.0.0', port=8000)
```
## tipsi_tools.drf.serializers.EnumSerializer
Allow you to deserealize incoming strings into `Enum` values.
You should add `EnumSerializer` into your serializers by hand.
```python
from enum import IntEnum
from django.db import models
from rest_framework import serializers
from tipsi_tools.drf.serializers import EnumSerializer
class MyEnum(IntEnum):
one = 1
two = 2
class ExampleModel(models.Model):
value = models.IntegerField(choices=[(x.name, x.value) for x in MyEnum])
class ExampleSerializer(serializers.ModelSerializer):
value = EnumSerializer(MyEnum)
# this allows you to post value as: {'value': 'one'}
```
Due to `Enum` and `IntegerField` realizations you may use `Enum.value` in querysets
```python
ExampleModel.objects.filter(value=MyEnum.two)
```
## tipsi_tools.django.log_requests.LoggerMiddleware
LoggerMiddleware will log request meta + raw post data into log.
For django<1.10 please use `tipsi_tools.django.log_requests.DeprecatedLoggerMiddleware`
## tipsi_tools.django.request_uniq
Decorator adds a unique for each uwsgi request dict as first function
argument.
For tests mock `_get_request_unique_cache`
## tipsi_tools.django.call_once_on_commit
Make function called only once on transaction commit. Here is examples
where function `do_some_useful` will be called only once after
transaction has been committed.
```python
class SomeModel(models.Model):
name = IntegerField()
@call_once_on_commit
def do_some_useful():
pass
def hook(sender, instance, **kwargs):
do_some_useful()
models.signals.post_save.connect(hook, sender=SomeModel)
with transaction.atomic():
some_model = SomeModel()
some_model.name = 'One'
some_model.save()
some_model.name = 'Two'
some_model.save()
```
For tests with nested transactions (commit actually most times is not
called) it is useful to override behaviour `call_once_on_commit`
when decorated function executed right in place where it is called.
To do so mock `on_commit` function. Example pytest fixture:
```
@pytest.fixture(scope='session', autouse=True)
def immediate_on_commit():
def side_effect():
return lambda f: f()
with mock.patch('tipsi_tools.django.on_commit', side_effect=side_effect) as m:
yield m
```
## tipsi_tools.django.fields.ChoicesEnum
Used for choices attribute for in model field
```
class FooBarEnum(ChoicesEnum):
foo = 1
bar = 2
class ExampleModel(models.Model):
type = models.IntegerField(choices=FooBarEnum.get_choices())
```
## tipsi_tools.django.db.utils.set_word_similarity_threshold
Allow to set postgres trigram word similarity threshold for default django database connection
```
set_word_similarity_threshold(0.4)
```
## tipsi_tools.django.db.pgfields.LTreeField
Django postgres ltree field type
```
class LTreeExample(models.Model):
path = LTreeField()
```
## tipsi_tools.django.db.pgfields.LTreeDescendant
Lookup for postgres ltree
```
# Add this import to models.py (file should be imported before lookup usage)
import tipsi_tools.django.db.pgfields # noqa
LTreeExample.objects.filter(path__descendant='root.level1')
```
## tipsi_tools.django.db.pgfields.SimilarityLookup
Postgres `text %> text` operator
```
# Add this import to models.py (file should be imported before lookup usage)
import tipsi_tools.django.db.pgfields # noqa
Books.objects.filter(title__similar='Animal Farm')
```
## tipsi_tools.django.db.pgfields.WordSimilarity
Postgres `text1 <<-> text2` operator. It returns `1 - word_similarity(text1, text2)`
```
from django.db.models import Value, F
similarity = WordSimilarity(Value('Animal Farm'), F('title'))
Books.objects.annotate(similarity=similarity)
```
## tipsi_tools.drf.use_form
Helps to use power of serializers for simple APIs checks.
```python
from rest_framework import serializers
from rest_framework.decorators import api_view
from tipsi_tools.drf import use_form
class SimpleForm(serializers.Serializer):
test_int = serializers.IntegerField()
test_str = serializers.CharField()
@api_view(['GET'])
@use_form(SimpleForm)
def my_api(data):
print(f'Data: {data["test_int"]} and {data["test_str"]}')
```
## tipsi_tools.aio_utils.DbRecordsProcessorWorker
Asyncio worker which wait for new records in postgres db table and process them.
## tipsi_tools.aio_utils.dict_query/sql_update
aiopg shortcuts
## tipsi_tools.python.execfile
Backport of python's 2 `execfile` function.
Usage: execfile('path/to/file.py', globals(), locals())
Returns: True if file exists and executed, False if file doesn't exist
## tipsi_tools.doc_utils.tipsi_sphinx
Sphinx extensions to generate documentation for django restframework serializers and examples for http requests.
In order to use them specify dependency for package installation:
```
pip install tipsi_tools[doc_utils]
```
Usage:
```
# Add to Sphinx conf.py
extensions = [
# ...
'tipsi_tools.doc_utils.tipsi_sphinx.dyn_serializer',
'tipsi_tools.doc_utils.tipsi_sphinx.http_log'
]
```
## Commands
### tipsi_env_yaml
Convert template yaml with substituion of `%{ENV_NAME}` strings to appropriate environment variables.
Usage: `tipsi_env_yaml src_file dst_file`
### tipsi_ci_script
Helper to run default CI pipeline. Defaults are set up for [giltab defaults](https://docs.gitlab.com/ee/ci/variables/#predefined-variables-environment-variables). Includes stages:
* build docker image with temporary name (commit sha by default)
* run tests (optional)
* push branch (by default only for master and staging branches)
* push tag, if there are tags
* cache image with common name
* delete image with temporary name
It's optimized for parallel launches, so you need to use unique temporary name (`--temp-name`). We want keep our system clean if possible, so we'll delete this tag in the end. But we don't want to repeat basic steps over and over, so we will cache image with common cache name (`--cache-name`), it will remove previous cached image.
### tipsi_wait
Wait for socket awailable/not-awailable with timeout.
```
# Wait until database port up for 180 seconds
tipsi_wait -t 180 postgres 5432
# Wait until nginx port down for 30 seconds
tipsi_wait -t 30 nginx 80
```
### run_filebeat
* checks environmental variables `-e KEY=VALUE -e KEY2=VALUE2`
* converts yaml template `tipsi_env_yaml {TEMPLATE} /tmp/filebeat.yml`
* run `/usr/bin/filebeat /tmp/filebeat.yml`
```
run_filebeat -e CHECKME=VALUE path_to_template
```
### doc_serializer
* output rst with list of serializers
* generates documentation artifacts for serializers
```
usage: doc_serializer [-h] [--rst] [--artifacts]
Parse serializers sources
optional arguments:
-h, --help show this help message and exit
--rst Output rst with serializers
--artifacts Write serializers artifacts
```
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
tipsi_tools-1.41.0.tar.gz
(33.4 kB
view hashes)