Skip to main content

an advanced dynamic task flow management on top of Celery

Project description

An advanced task flow management on top of Celery.

codecov PyPI Current Version PyPI Implementation PyPI Wheel Travis CI Documentation Status GitHub stars GitHub license Twitter

Is this project helpful? Send me a simple warm message!

Crossroad

Last stable release: Selinon 1.3.0

TLDR;

An advanced flow management above Celery (an asynchronous distributed task queue) written in Python3, that allows you to:

  • Dynamically schedule tasks based on results of previous tasks

  • Group tasks into flows in simple YAML configuration files

  • Schedule flows from other flows (even recursively)

  • Store results of tasks in your storages and databases transparently, validate results against defined JSON schemas

  • Do redeployment respecting changes in the YAML configuration files without purging queues (migrations)

  • Track flow progress via the build-in tracing mechanism

  • Complex per-task or per-flow failure handling with fallback tasks or fallback flows

  • No DAG limitation in your flows

  • Selectively pick tasks in your flow graphs that should be executed respecting task dependencies

  • Make your deployment easy to orchestrate using orchestration tools such as Kubernetes

  • Highly scalable Turing complete solution for big data processing pipelines

  • And (of course) much more… check docs

YouTube Video

Let’s explain Selinon using a YouTube video (click to redirect to YouTube).

Ironwoodland

About

This tool is an implementation above Celery that enables you to define flows and dependencies in flows, schedule tasks based on results of Celery workers, their success or any external events. If you are not familiar with Celery, check out its homepage www.celeryproject.org or this nice tutorial.

Selinon was originally designed to take care of advanced flows in one of Red Hat products, where it already served thousands of flows and tasks. Its main aim is to simplify specifying group of tasks, grouping tasks into flows, handle data and execution dependencies between tasks and flows, easily reuse tasks and flows, model advanced execution units in YAML configuration files and make the whole system easy to model, easy to maintain and easy to debug.

By placing declarative configuration of the whole system into YAML files you can keep tasks as simple as needed. Storing results of tasks in databases, modeling dependencies or executing fallback tasks/flows on failures are separated from task logic. This gives you a power to dynamically change task and flow dependencies on demand, optimize data retrieval and data storage from databases per task bases or even track progress based on events traced in the system.

Selinon was designed to serve millions of tasks in clusters or data centers orchestrated by Kubernetes, OpenShift or any other orchestration tool, but can simplify even small systems. Moreover, Selinon can make them easily scalable in the future and make developer’s life much easier.

A Quick First Overview

Selinon is serving recipes in a distributed environment, so let’s make a dinner!

If we want to make a dinner, we need to buy ingredients. These ingredients are bought in buyIngredientsFlow. This flow consists of multiple tasks, but let’s focus on our main flow. Once all ingredients are bought, we can start preparing our dinner in prepareFlow. Again, this flow consists of some additional steps that need to be done in order to accomplish our future needs. As you can see, if anything goes wrong in mentioned flows (see red arrows), we make a fallback to pizza with beer which we order. To make beer cool, we place it to our Fridge storage. If we successfully finished prepareFlow after successful shopping, we can proceed to serveDinnerFlow.

Just to point out - grey nodes represent flows (which can be made of other flows or tasks) and white (rounded) nodes are tasks. Conditions are represented in hexagons (see bellow). Black arrows represent time or data dependencies between our nodes, grey arrows pinpoint where results of tasks are stored.

Main dinner flow

For our dinner we need eggs, flour and some additional ingredients. Moreover, we conditionally buy a flower based on our condition. Our task BuyFlowerTask will not be scheduled (or executed) if our condition is False. Conditions are made of predicates and these predicates can be grouped as desired with logical operators. You can define your own predicates if you want (default are available in selinon.predicates). Everything that is bought is stored in Basket storage transparently.

Let’s visualise our buyIngredientsFlow:

Buy ingredients flow

As stated in our main flow after buying ingredients, we proceed to dinner preparation but first we need to check our recipe that is hosted at http://recipes.lan/how-to-bake-pie.html. Any ingredients we bought are transparently retrieved from defined storage as defined in our YAML configuration file. We warm up our oven to expected temperature and once the temperature is reached and we have finished with dough, we can proceed to baking.

Based on the description above, our prepareFlow will look like the following graph:

Prepare flow

Once everything is done we serve plates. As we want to serve plates for all guests we need to make sure we schedule N tasks of type ServePlateTask. Each time we run our whole dinner flow, number of guests may vary so make sure no guest stays hungry. Our serveDinnerFlow would look like the following graph:

Serve dinner flow

This example demonstrates very simple flows. The whole configuration can be found here. Just check it out how you can easily define your flows! You can find a script that visualises graphs based on the YAML configuration in this repo as well.

More info

The example was intentionally simplified. You can also parametrize your flows, schedule N tasks (where N is a run-time evaluated variable), do result caching, placing tasks on separate queues in order to be capable of doing fluent system updates, throttle execution of certain tasks in time, propagate results of tasks to sub-flows etc. Just check documentation for more info.

Live Demo

A live demo with few examples can be found here. Feel free to check it out.

Installation

$ pip3 install selinon

Available extras:

Extras can be installed via:

$ pip3 install selinon[celery,mongodb,postgresql,redis,s3,sentry]

Feel free to select only needed extras for your setup.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

selinon-1.3.0.post0.tar.gz (215.5 kB view details)

Uploaded Source

Built Distribution

selinon-1.3.0.post0-py3-none-any.whl (324.5 kB view details)

Uploaded Python 3

File details

Details for the file selinon-1.3.0.post0.tar.gz.

File metadata

  • Download URL: selinon-1.3.0.post0.tar.gz
  • Upload date:
  • Size: 215.5 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.1 CPython/3.9.7

File hashes

Hashes for selinon-1.3.0.post0.tar.gz
Algorithm Hash digest
SHA256 926010ccf6853800addd2dcbeb367ad39f97e8f1c3c451d850427239d6f05f53
MD5 2c8de57dc1c887829abfb500ca0ff991
BLAKE2b-256 320aa1fc9640ab9598800d2ed964e01e157897818efcb9fa8f1d93dd74c31e5b

See more details on using hashes here.

File details

Details for the file selinon-1.3.0.post0-py3-none-any.whl.

File metadata

File hashes

Hashes for selinon-1.3.0.post0-py3-none-any.whl
Algorithm Hash digest
SHA256 7fc703ed791b55cacf6440d7cc5c192bf938971d6176512fb6b202ca68a1ba99
MD5 a7fc3ec50d0857d369dcab32e3880b34
BLAKE2b-256 985aead623118ccd4efeb4c9b8750afdf7796d48afbd4ae971a1c35d066a6d3e

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page