Skip to main content

Python package for starflow.

Project description

Build amazing ideas, piece by piece.

Docs DOI


create-workflow

Table of contents


About

starflow is an open source workflow management platform, with:

  • :desktop_computer: an intuitive Graphical User Interface that facilitates creating, editing and monitoring any type of Workflow, from data processing to machine learning
  • :package: a standard way of writing and publishing functional Pieces, which follows good practices for data modeling, documentation and distribution
  • :gear: a REST API that controls a running Apache Airflow instance

Creating Workflows in the GUI is as simple as dragging and dropping Pieces to the canvas, and connecting them. The user can schedule the Workflow to run periodically, at a specific date/time, or trigger it manually. The monitoring page shows the status of each Workflow Piece in real time, including the logs and results of each run.

Pieces are functional units that can be reused in multiple Workflows. Pieces can execute anything that can be written in Python, and can be easily distributed and installed directly from Github repositories to be used in starflow Workflows.

Every starflow Workflow corresponds to an Apache Airflow DAG, and each Piece corresponds to an Airflow task. starflow controls an Airflow instance, which is responsible for executing, scheduling and monitoring the Workflows (DAGs).

You can think of starflow as Airflow with superpowers:

  • :desktop_computer: create highly complex Workflows with simple point-and-click and drag-and-drop operations, in an user-friendly GUI
  • :package: make use of Pieces developed by other people, share and reuse your own Pieces
  • :busts_in_silhouette: collaborate in groups to edit and monitor Workflows
  • :chart_with_upwards_trend: experience a cleaner and more intuitive GUI for viewing Workflows results, including logs and richer reports with images and tables
  • :minidisc: shared storage for tasks in the same workflow
  • :arrows_counterclockwise: use gitSync to sync DAGs from files stored in a Git repository
  • :wheel_of_dharma: scalable, Kubernetes-native platform
  • :battery: powered by Apache Airflow for top-tier workflows scheduling and monitoring

Quick start

Check out the quick start guide in the documentation.

The starflow Python package can be installed via pip. We reccommend you install starflow in a separate Python environment.

pip install starflow-py[cli]

You can then use starflow command line interface to easily run the starflow platform locally (requires Docker Compose V2). Go to a new, empty directory and run the following command:

starflow platform run-compose

After all processes started successfully, navigate to localhost:3000 to access the starflow frontend service.
Obs.: the first time you run the platform, it may take a few minutes to download the Docker images.

Running the starflow platform locally with Docker compose is useful for development and testing purposes. For production environments, we recommend you deploy starflow and Airflow to a Kubernetes cluster. For other deployment modes, check out the instructions in the documentation.


GUI

The starflow frontend service is a React application that provides the GUI for easily creating, editing and monitoring Workflows. Check out the GUI documentation for more details.

Access authentication Sign up and login to use the starflow platform.

signup and login

Select or Create Workspaces Select an existing or create a new Workspace.

create workspace

Install Pieces repositories Install bundles of Pieces to your starflow Workspaces direclty from Github repositories, and use them in your Workflows.

install pieces

Create Workflows Create Workflows by dragging and dropping Pieces to the canvas, and connecting them.

create-workflow

Edit Pieces Edit Pieces by changing their input. Outputs from upstream Pieces are automatically available as inputs for downstream Pieces. Pieces can pass forward any type of data, from simple strings to heavy files, all automatically handled by starflow shared storage system.

edit pieces

Configure Workflows Configure and schedule Workflows to run periodically, at a specific date/time, or trigger them manually.

schedule workflows

Monitor Workflows Monitor Workflows in real time, including the status of each Piece, the logs and results of each run.

run-pieces7


Pieces

Pieces

Pieces are the secret sauce of starflow, they are functional units that can be distributed and reused in multiple Workflows. starflow Pieces are special because they:

  • :snake: can execute anything written in Python, heavy-weight (e.g. Machine Learning) as well as light-weight (e.g. sending emails) tasks
  • :traffic_light: have well defined data models for inputs, outputs and secrets
  • :package: run in self-contained and isolated execution environments (Docker containers)
  • :gear: are immutable, guaranteeing reproducibility of your workflows
  • :octocat: are organized in git repositories, for easy packaging, distribution and installation
  • :bookmark_tabs: are properly versioned, tested and documented
  • :zap: are plug-and-play and versatile, can be easily incorporated in any workflow

It is very easy to create and share your own Pieces:

  1. write your Python function as a Piece
  2. define the data types, dependencies, metadata and tests
  3. publish in a git repository (public or private)

The Pieces repository template provides the basic structure, example files and automatic actions for a seamless Pieces creation experience.

Read more in the Pieces documentation.


REST

The Backend service is a REST API that controls a running Apache Airflow instance. It is responsible for:

  • executing operations requested by the frontend service
  • interacting with the Airflow instance, including triggering, creating, editing and deleting Workflows (DAGs)
  • interacting with the starflow Database

The REST service is written in Python, using the FastAPI framework. Read more about it in the REST documentation.


Credits

starflow is developed and maintained by [Prochain].

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

starflow_py-0.47.0.tar.gz (76.6 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

starflow_py-0.47.0-py3-none-any.whl (99.3 kB view details)

Uploaded Python 3

File details

Details for the file starflow_py-0.47.0.tar.gz.

File metadata

  • Download URL: starflow_py-0.47.0.tar.gz
  • Upload date:
  • Size: 76.6 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.8.18

File hashes

Hashes for starflow_py-0.47.0.tar.gz
Algorithm Hash digest
SHA256 e3c958d9b5e4553ef9d606481261b1a95e6dc4f748e27aa8435bb71b91274a73
MD5 299a31156769fe2095db21afee053426
BLAKE2b-256 0f4ed69c8cfa60f9bf87ef9e348d12f9a98c35095c922ca7d257449710b9d8c6

See more details on using hashes here.

File details

Details for the file starflow_py-0.47.0-py3-none-any.whl.

File metadata

  • Download URL: starflow_py-0.47.0-py3-none-any.whl
  • Upload date:
  • Size: 99.3 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.8.18

File hashes

Hashes for starflow_py-0.47.0-py3-none-any.whl
Algorithm Hash digest
SHA256 b736e3b704857138b706003cf45bd4d062c6445439f92371a4ba06893b7a6843
MD5 f859635b04c147e875d6bf00508837d3
BLAKE2b-256 a27f8bd6fab8293f1d528d2619ed2127524063518d7dfe91403943d5ed0ec9b3

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page