Skip to main content

Run FireWorks workflows in Google Cloud

Project description

Borealis

Runs FireWorks workflows on Google Compute Engine (GCE).

See the repo Borealis and the PyPI page borealis-fireworks.

  • Borealis is the git repo name.
  • borealis-fireworks is the PyPI package name.
  • borealis-fireworker.service is the name of the systemd service.
  • fireworker is the recommended process username and home directory name.

What is it?

FireWorks is open-source software for defining, managing, and executing workflows. Among the many workflow systems, FireWorks is exceptionally straightforward, lightweight, and adaptable. It's well tested and supported. The only shared services it needs are a MongoDB server (acting as the workflow "LaunchPad") and a file store.

Borealis lets you spin up as many temporary worker machines as you want in the Google Cloud Platform to run your workflow. That means pay-per-use and no contention between workflows.

How does Borealis support workflows on Google Compute Engine?

TL;DR: Spin up worker machines when you need them, deploy your task code to the workers in Docker Images, and store the data in Google Cloud Storage instead of NFS.

Worker VMs: As a cloud computing platform, Google Compute Engine (GCE) has a vast number of machines available. You can spin up lots of GCE "instances" (also called Virtual Machines or VMs) to run your workflow, change your code, re-run some tasks, then let the workers time out and shut down. Google will charge you based on usage and there's no resource contention between workflows.

Borealis provides the ComputeEngine class and its command line wrapper gce to create, tweak, and delete groups of worker VMs.

Borealis provides the fireworker Python script to run as the top level program for each worker. It's a wrapper around FireWorks' rlaunch feature.

You can run these Fireworkers on and off GCE, or both at the same time, as long as all the workers can connect to your FireWorks "LaunchPad" server and the data store.

Docker: You need to deploy your "payload" task code to those GCE VMs. If it's Python source code, it needs a particular runtime environment: Python 2.7 or 3.something, Python pip packages, Linux apt packages, compiled Cython code, data files, and environment variable settings. A GCE VM starts up from a GCE Disk Image which could have all that preinstalled (with or without the Python source code) but it'd be hard to keep it up to date, hard to share it with your team, and hard to keep track of how to reproduce it.

This is what Docker Images are designed for. You maintain a Dockerfile containing repeatable instructions to build your payload Image, then run docker locally or use the Google Cloud Build service to build the Image and store it in the Google Container Registry.

Borealis provides the DockerTask Firetask to run just such a payload. It pulls a named Docker Image, starts up a Docker Container, runs a given shell command in that Container, and shuts down the Container. Running within Docker also isolates the task's runtime environment and side effects from other tasks and the Fireworker.

DockerTask logs the Container's stdout + stderr to a file and to Python logging (which fireworker connects to StackDriver). DockerTask also imposes a given timeout on the task.

Google Cloud Storage: Although you can set up an NFS shared file service for the workers' files, Google Cloud Storage (GCS) is the native storage service. GCS costs literally 1/10th as much as NFS service and it scales up better. GCS lets you archive your files in yet lower cost tiers intended for infrequent access. Pretty much all of Google's cloud services revolve around GCS, e.g., Pub/Sub can trigger an action on a particular upload to GCS.

But Cloud Storage is not a file system. It's an object store with a lighter weight protocol to fetch/store/list whole files, called "blobs." It does not support simultaneous writers, rather, the last "store" of a blob wins. Blob pathnames can contain / characters but GCS doesn't have actual directory objects, so e.g. there's no way to atomically rename a directory.

DockerTask supports Cloud Storage by fetching the task's input files from GCS and storing its output files to GCS.

You can access your GCS data via the gsutil command line tool, the gcsfuse mounting tool, and the Storage Browser in the Google Cloud Platform web console.

Logging: fireworker sets up Python logging to write to Google's StackDriver logging service so you can watch all your workers running in real time.

Projects: With Google Cloud Platform, you set up a project for your team to use. All services, VMs, data, and access controls are scoped by the project.

Borealis Components

gce: Borealis provides the ComputeEngine class and its command line wrapper gce to create, tweak, and delete a group of worker VMs. ComputeEngine will pass in the needed launch parameters such as connection details to access the LaunchPad server. After you generate a workflow description (DAG), call FireWorks' LaunchPad.add_wf() method or run FireWorks' lpad add command line tool to upload it to the LaunchPad. Then you can call the ComputeEngine.create() method or the gce command line to launch a bunch of worker VMs to run the workflow.

ComputeEngine and gce are also useful for immediately deleting a batch of worker VMs or asking them to quit cleanly between Firetasks, although on their own they'll shut down after an idle timeout. The idle timeout duration is longer if there are queued tasks that are waiting on other tasks to finish.

ComputeEngine and gce can also set GCE metadata fields on a batch of workers. This is used to implement the --quit-soon feature.

fireworker: Borealis provides the fireworker Python script to run as a worker. fireworker gets the worker launch parameters and calls the FireWorks library to "rapidfire" launch your FireWorks "rockets." It also handles server shutdown.

fireworker sets up Python logging and connects it to Google Cloud's StackDriver logging so you can watch all your worker machines in real time.

To run fireworker on GCE VMs, you'll need to create a GCE Disk Image that contains Python, the borealis-fireworks pip, and such. See the instructions in how-to-install-gce-server.txt.

The fireworker command can also run on your local computer for easier debugging. For that, you'll need to install the borealis-fireworks pip and set up your computer to access the right Google Cloud Project.

TODO: Document how to set up access to the GCP project.

DockerTask: The DockerTask Firetask will pull a named Docker Image, start up a Docker Container, run a given shell command in that Container, and stop the container. This is a reliable way to deploy your payload code packaged up with its runtime environment to the workers. It also isolates the payload from the Fireworker and from all other Firetasks.

Docker always runs a shell command in the Container. If you want to run a Firetask in the Container, include a little Python script to bridge the gap: It takes a Firetask name and a JSON dictionary as command line arguments, instantiates the Firetask with those arguments, and calls the Firetask's run_task() method.

DockerTask supports Google Cloud Storage (GCS) by fetching the task's input files from GCS, mapping it into the Docker Container, and storing the task's output files to GCS. This requires you to declare the input and output paths. (With these declarations, a workflow builder can compute the task interdependencies that FireWorks needs.) Any path ending with a / denotes a directory tree of files.

When storing task outputs, DockerTask creates blobs with names ending in / that act as "directory placeholders" to speed up tree-oriented list-blob requests. This means you can run gcsfuse without using the --implicit-dirs flag, resulting in mounted directories that run 10x faster.

DockerTask imposes a given timeout on the shell command so it can't run forever.

DockerTask logs the Container's stdout and stderr to a file and to Python logging (which fireworker connects to StackDriver).

Team Setup

TODO: Install & configure dev tools, create a GCP project, auth stuff, install MongoDB on a GCE VM or set up Google-managed MongoDB, create a Fireworker disk image & image family, ...

See borealis/setup/how-to-install-gce-server.txt for detail instructions to set up your Compute Engine Disk Image and its "Service Account" for authorization.

xxxxx to connect to the LaunchPad MongoDB server. Metadata parameters and the worker's my_launchpad.yaml file configure the Fireworker's MongoDB host, port, DB name, and idle timeout durations. Users can have their own DB names on a shared MongoDB server, and each user can have multiple DB names -- each an independent launchpad space for workflows and their Fireworker nodes.

Individual Developer Setup

TODO: Install & configure dev tools, make a storage bucket with a globally-unique name, build a Docker image to run, ...

Run

TODO

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

borealis-fireworks-0.5.1.tar.gz (35.4 kB view details)

Uploaded Source

Built Distribution

borealis_fireworks-0.5.1-py2.py3-none-any.whl (35.0 kB view details)

Uploaded Python 2 Python 3

File details

Details for the file borealis-fireworks-0.5.1.tar.gz.

File metadata

  • Download URL: borealis-fireworks-0.5.1.tar.gz
  • Upload date:
  • Size: 35.4 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.1.1 pkginfo/1.5.0.1 requests/2.23.0 setuptools/45.2.0 requests-toolbelt/0.9.1 tqdm/4.42.1 CPython/3.8.0

File hashes

Hashes for borealis-fireworks-0.5.1.tar.gz
Algorithm Hash digest
SHA256 968dc06cae3d8338708ee18c141321b815d7329b5f6864a73bdeb422bc24c3c3
MD5 a8ddf601a131465e1b01da908bf7aa22
BLAKE2b-256 136b7ab02e540625b91d083bc8954f747bea6900c181d1d2fb8d0f083db160c4

See more details on using hashes here.

File details

Details for the file borealis_fireworks-0.5.1-py2.py3-none-any.whl.

File metadata

  • Download URL: borealis_fireworks-0.5.1-py2.py3-none-any.whl
  • Upload date:
  • Size: 35.0 kB
  • Tags: Python 2, Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.1.1 pkginfo/1.5.0.1 requests/2.23.0 setuptools/45.2.0 requests-toolbelt/0.9.1 tqdm/4.42.1 CPython/3.8.0

File hashes

Hashes for borealis_fireworks-0.5.1-py2.py3-none-any.whl
Algorithm Hash digest
SHA256 be44bf1cfd335c03159d25247e65281514abe318f59ddd3860f099bffdec4fb8
MD5 a8227b72c66889ea14195cbd9759f26a
BLAKE2b-256 f294345d241284b31246ff3b25dab9f3dbbd2de3a395891e2ab07f848204f0ed

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page