data linter
Project description
Data Linter
A python package to to allow automatic validation of data as part of a Data Engineering pipeline. It is designed to automate the process of moving data from Land to Raw-History as described in the ETL pipline guide
The validation is based on the goodtables
package, from the fine folk at Frictionless Data. More information can be found at their website.
Installation
pip install data_linter
Usage
This package takes a yaml
based config file written by the user (see example below), and validates data in the specified Land bucket against specified metadata. If the data conforms to the metadata, it is moved to the specified Raw bucket for the next step in the pipeline. Any failed checks are passed to a separate bucket for testing. The package also generates logs to allow you to explore issues in more detail.
To run the validation, at most simple you can use the following:
from data_linter import run_validation
config_path = "config.yaml"
run_validation(config_path)
Example config file
land-base-path: s3://land-bucket/my-folder/ # Where to get the data from
fail-base-path: s3://fail-bucket/my-folder/ # Where to write the data if failed
pass-base-path: s3://pass-bucket/my-folder/ # Where to write the data if passed
log-base-path: s3://log-bucket/my-folder/ # Where to write logs
compress-data: true # Compress data when moving elsewhere
remove-tables-on-pass: true # Delete the tables in land if validation passes
all-must-pass: true # Only move data if all tables have passed
fail-unknown-files:
exceptions:
- additional_file.txt
- another_additional_file.txt
# Tables to validate
tables:
table1:
required: true # Does the table have to exist
pattern: null # Assumes file is called table1
metadata: meta_data/table1.json
linter: goodtables
table2:
required: true
pattern: ^table2
metadata: meta_data/table2.json
How to update
We have tests that run on the current state of the poetry.lock
file (i.e. the current dependencies). We also run tests based on the most up to date dependencies allowed in pyproject.toml
. This allows us to see if there will be any issues when updating dependences. These can be run locally in the tests
folder.
When updating this package, make sure to change the version number in pyproject.toml
and describe the change in CHANGELOG.md.
If you have changed any dependencies in pyproject.toml
, run poetry update
to update poetry.lock
.
Once you have created a release in GitHub, to publish the latest version to PyPI, run:
poetry build
poetry publish -u <username>
Here, you should substitute for your PyPI username. In order to publish to PyPI, you must be an owner of the project.
Process Diagram
How logic works
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file data_linter-1.1.1.tar.gz
.
File metadata
- Download URL: data_linter-1.1.1.tar.gz
- Upload date:
- Size: 11.9 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: poetry/1.0.10 CPython/3.8.5 Darwin/19.6.0
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | c152a161815bd8a4b8d0f113ba51c4c23853fbc4f26f853e79692aecb024597d |
|
MD5 | b93921f4965b6f84df42fceceff58d75 |
|
BLAKE2b-256 | d9a25a03856702e8d41456fa097c51e8eab7decef00bfa13dc2e17f3561d4adf |
File details
Details for the file data_linter-1.1.1-py3-none-any.whl
.
File metadata
- Download URL: data_linter-1.1.1-py3-none-any.whl
- Upload date:
- Size: 12.4 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: poetry/1.0.10 CPython/3.8.5 Darwin/19.6.0
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 55796a9b68842951a77c5f1c119929b1c92aa689f76e0c1d4e05409485041ab5 |
|
MD5 | f1c68107ea88b4e23d00e48bf08810c4 |
|
BLAKE2b-256 | 1a0ea71e8733b45509ae399f9481566b447caa195a641b48bedf404ff3c12854 |