A fluent API layer for tensorflow extended e2e machine learning pipelines
Project description
fluent-tfx
Fluent TFX provides a better API for TFX pipeline creation. If you are already using tensorflow or keras for your models, this is an easy to use api that will have your model up and running in an e2e scenario quickly, without the need to waste a significant amount of time of learning the internals of tfx to get something working.
pip install fluent-tfx
import fluent_tfx as ftfx
Goals of this package
Create an fluent, concise and easy to use API for TFX pipeline composition.
Usage
This is a lightweight api to aid with the construction of tfx pipelines. Every side-effect and produced artifact is 100% compatible with the rest of the tfx sybsystem, including but not limited to: all the supported Beam runners (ex. Local with BeamDagRunner
, Kubeflow, Airflow, Dataflow, etc.), custom components, ML Metadata artifacts.
Please take a look into examples/usage_guide
for a full walkthrough of the provided pipeline construction tools.
It provides several shortcut functions and utilities for easier, more readable, compact and expressive pipeline definitions, with sensible defaults. Some assumptions are made about specific inputs and outputs of the componens which are described after this small example:
This is what you need to get started by using fluent tfx, instead of ~ 20 files produced with the tfx template cli commands:
Pipeline Creation
# file pipeline.py
import fluent_tfx as ftfx
def get_pipeline():
return ftfx.PipelineDef(name='taxi_pipeline') \
.with_sqlite_ml_metadata() \ # or provide a different configuration in the constructor
.from_csv(uri='./examples/taxi/data') \ # or use bigquery/tfrecord/custom components
.generate_statistics() \ # or not (optional)
.infer_schema() \ # or use with_imported_schema(<uri>) to load your schema and detect anomalies
.preprocess(<preprocessing_fn>) \
.tune(<tune_args>) \ # or use with_hyperparameters(<uri>) to import a best set of hyperparameters straight to the model--or skip tuning and just use constants on .train()
.train(<trainer_fn>, <train_args>) \
.evaluate_model(<eval_args>) \ # evaluate against baseline and bless model
.infra_validate(<args>) \
.push_to(<pusher_args>) \
.cache() \ # optional
.with_beam_pipeline_args(<args>) \ # optional too
.build()
# run normally with:
# python -m pipeline
if __name__ == '__main__':
pipeline = get_pipeline()
BeamDagRunner().run(pipeline)
Utilities
Input Builders
Even if you don't use PipelineDef
for your pipeline creation, you can still use the 1-liner methods for input creation, found under ftfx.input_builders
. Including but not limited to: from_csv
, from_tfrecord
, with_hyperparameters
, get_latest_blessed_model_resolver
.
Other
WIP
Assumptions and Degrees of Freedom
Custom components are supported to a large extent, but there will still some edge cases that would only work with the verbose plain old tfx api.
Assumptions are related to component dag wiring, paths and naming.
Paths
PipelineDef
needs apipeline_name
and an optionalbucket
path.- Binary/Temporary/Staging artifacts are stored under
{bucket}/{name}/staging
- Default ml metadata sqlite path is set to
{bucket}/{name}/metadata.db
unless specified otherwise bucket
defaults to./bucket
- Pusher's
relative_push_uri
will publish the model to{bucket}/{name}/{relative_push_uri}
Component IO and Names
- An input, or an
example_gen
component provides.tfrecord
s (probably in gzipped format) to next components - Fluent TFX follows the TFX naming of default components for everything. When providing custom components, make sure that inputs and outputs are on par with TFX.
- For example, your custom
example_gen
component should have a.outputs['examples']
attribute - When using extra components from
input_builders
make sure that the names you provide not override defaults, such as standard tfx component names assnake_case
and{name}_examples_provider
,user_{x}_importer
.
Component Wiring Defaults
- If a user provided schema uri is provided, it will be used for data validation, transform, etc. The generated schema component will still generate artifacts if declared
- If a user did not provide a model evaluation step, it will not be wired to the pusher
- The default training input source are transform outputs. The user can specify if he wants the raw tf records instead
- If hyperparameters are specified and tuner is declared, tuner will still test configurations and produce hyperparameter artifacts, but the provided hyperparameters will be used
Runners
There is no extra effort required to run the pipeline on different runners, nor extra dependencies required: PipelineDef
produces a vanilla tfx pipeline.
However, if you are using ftfx
utilities inside your pipeline functions, be sure to include this package in your requirements.txt beam argument
Examples
Example usages and templated quickstarts are available under <repo root>/examples
.
What made me create this fluent api
- The TFX api is aimed for maximum flexibility. A very big portion of machine learning pipelines can be created with a much less verbose and ultimately, simpler api.
- By the time of creation (mid 2020), the TFX demos are all over the place, with deprecated usage in many places, the documentation is lacking and there are a lot of issues that need fixing.
- The default scaffolding is horrible: On the one hand, it makes it easy to get started, but on the other hand, it creates 10~20 files that make it hard to keep track of everything even if you are not new to this kind of engineering.
- Why use scaffolding anyway? The default TFX api is flexible as stated above, but there is (1) too much magic going on and (2) lot's of components IOs could be routed automatically.
- The pipeline definition is simply too much LoC for no apparent reason and the examples are lengthy, making it hard to keep track of everything
What this package contains/is going to do
- Provide an easy to use, fluent API for configuration, instead of scaffolding huge directory structures
- Support all the runners and the functionality of tfx
- Use as much code from tfx as possible, including components, internally
- Keep the pb2 (protocol buffers) on the extenral api because they are a powerful tool
- Keep usage restrictions to a sensible minimum, while enabling users to specify custom components and logic wherever possible
- Provide utility functions and components for some tfx edge cases and boilerplate code
- Support custom user-provided components
What this package is not
First of all, it is not a non-opinionated way of making machine learning end to end pipelines. This does not automatically solve all data engineering and good statistical practises problems There is going to be no estimator support for anything, sorry. Please migrate to new, native keras with tensorflow 2.0.
Please be advised, this project is aimed to make the majority of machine learning deployment to production tasks easier. Some parts of the API might take an opinionated approach to specific problems.
If you've got suggestions for improvement, please create a new issue and we can chat about it :).
But Tensorflow Extended is already fully capable to construct e2e pipelines by itself, why bother to use another API?
- Verbose and long code definitions. Actual preprocessing and training code can be as lengthy as an actual pipeline component definition.
- Lack of sensible defaults. You have to manually specify inputs and outputs to everything. This allows maximum flexibility on one hand, but on the other 99% of cases, most of the IOs can be automatically wired. For example, your preprocessing component is going to probably read your first input component's input, and pass outputs to training.
- Too much boilerplate code. Scaffolding via the TFX CLI produces 15–20 files in 4–5 directories.
The benefits of an easier to use, API layer
- Fluent and compact pipeline definition and runtime configuration. No more scrolling through endless, huge 300+ line functions that construct pipelines
- No scaffolding, easy to set up by using a few lines of code
- Extra helpful utilities to speed up common tasks, such as data input, TFX component construction and wiring
- Sensible defaults and 99% - suitable component IO wiring
Licensing Notice
Main source code of this package, under the directory fluent_tfx
is released under the MIT License.
Every example under the directory /examples
that is based on Google's ones is released under the APACHE 2.0 license, copyrighted by TFX Authors and Google LLC, with any modified parts being stated at the top of the notice on each file.
Every other example that is built from scratch is also, released under the MIT License.
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file fluent-tfx-0.23.1.tar.gz
.
File metadata
- Download URL: fluent-tfx-0.23.1.tar.gz
- Upload date:
- Size: 31.9 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/3.2.0 pkginfo/1.5.0.1 requests/2.24.0 setuptools/47.1.0 requests-toolbelt/0.9.1 tqdm/4.48.2 CPython/3.8.5
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 3b5c6c4bac2afab130dea403e687de4c7291e0ab35ab2e5774d9ca83ce97a6ca |
|
MD5 | 8dec64ffdddd9768a853144e485fb1bd |
|
BLAKE2b-256 | d3a64c5a0d2f296b2a26f5e83c14ffb0c716af3243f4fdc268afdca62248d787 |
File details
Details for the file fluent_tfx-0.23.1-py3-none-any.whl
.
File metadata
- Download URL: fluent_tfx-0.23.1-py3-none-any.whl
- Upload date:
- Size: 43.6 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/3.2.0 pkginfo/1.5.0.1 requests/2.24.0 setuptools/47.1.0 requests-toolbelt/0.9.1 tqdm/4.48.2 CPython/3.8.5
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 5d7101271a5c1d9d1489957072112fea4ad9c79f61bea46b9dc5be7f4fe6089f |
|
MD5 | bf588cb3af21ce26a1f4a683622a95cc |
|
BLAKE2b-256 | 2dce53fc379377210863ff95a02d6b56d445f582f36ee4df612bb4204d22639b |