Skip to main content

Execute workflows on Galaxy through the CLI

Project description

PyPI version fury.io Build Status

Galaxy workflow executor 0.2.6

This setup uses bioblend (0.12 - 0.13 tested) to run a Galaxy workflow through the CLI:

  • Inputs:
    • Galaxy workflow with steps annotated with labels as JSON file (MUST be obtained in Galaxy UI from Share Workflow -> Download).
    • Parameters dictionary as YAML (JSON also supported). Supports both simple input parameters and tools parameters not exposed by simple input parameters.
    • Input files specified as paths or dataset IDs in a YAML file.
    • Steps with allowed errors specified in a YAML file (optional)
    • Name for a history to be created (optional)

Galaxy workflow

The workflow should be annotated with labels, ideally for all steps, but at least for the steps where you want to be able to set parameters through the parameters dictionary. It should be the JSON file resulting from Workflows (upper menu) -> Share workflow (on the drop down menu of the workflow, in the workflow list) -> Download (in the following screen).

Execution state

The setup will store the execution state during the run, so that if there are disconnection or errors, it can restart following the progress of the same workflow. It stores the state by default in the working directory, in exec_state.pickle. This might not be session proof: having a failure running workflow A, then trying to run a subsequent workflow B you might get errors. So when switching running workflows, make sure to use either different working directories or specify the path to the state path explicitly through --state-file. Please note that to specify this for a new run, the file is not expected to exist.

The state file is deleted automatically on a successful execution.

Parameters YAML

The parameters YAML file can be generated for a desired workflow by executing:

generate_params_from_workflow.py -C galaxy_credentials.yaml \
                            -G test_instance -o test \
                            -W wf.json
  • Inputs:
    • Credentials file to a Galaxy instance (this file uses the same format as the one used by parsec)
    • Name of the Galaxy instance among those listed in the credentials file (optional).
    • Galaxy workflow as JSON file (from share workflow -> download)
    • Output directory path (optional)

The output wf-parameters.yaml will follow the following structure:

step_label_x:
   param_name: "value"
    ....
   nested_param_name:
        n_param_name: "n_value"
        ....
        x_param_name: "x_value"
step_label_x2:
    ....
....
other_galaxy_setup_params: { ... }

Input files in YAML

It should point to the files in the file system, set a name (which needs to match with a workflow input label) and file type (among those recognized by Galaxy).

The structure of the YAML file for inputs is:

matrix:
  path: /path/to/E-MTAB-4850.aggregated_filtered_counts.mtx
  type: txt
genes:
  path: /path/to/E-MTAB-4850.aggregated_filtered_counts.mtx_rows
  type: tsv
barcodes:
  path: /path/to/E-MTAB-4850.aggregated_filtered_counts.mtx_cols
  type: tsv
gtf:
  dataset_id: fe139k21xsak

where in this example case the Galaxy workflow should have input labels called matrix, genes, barcodes and gtf. The paths need to exist in the local file system, if path is set within an input. Alternatively to a path in the local file system, if the file is already on the Galaxy instance, the dataset_id of the file can be given instead, as shown for the gtf case here.

Steps with allowed errors

This optional YAML file indicates the executor which steps are allowed to fail without the overal execution being considered failed and hence retrieving result files anyway. This is to make room to the fact that on a production setup, there might be border conditions on datasets that could produce acceptable failures.

The structure of the file relies on the labels for steps used in the workflow and parameters files

step_label_x:
  - any
step_label_z:
  - 1
  - 43

The above example means that the step with label step_label_x can fail with any error code, whereas step with label step_label_z will only be allowed to fail with codes 1 or 43 (specific error code handling is not yet implemented).

Results

All workflow outputs that were marked in the workflow to be shown will be downloaded to the specified results directory, hidden results will be ignored. Unless specified, histories (with its contents) and workflows will be deleted from the instance.

Toy example

A simple example, which is used in the CI testing, can be seen and run locally through the run_tests_with_containers.sh script.

Exit error codes

Currently produced error codes:

Error code Description
3 Connection error during history deletion, this is not a critical error as most probably the history will get deleted by the server. A file named histories_to_check.txt is created in the working directory. Data will have been downloaded by then.
4 Workflow scheduling cancelled at the Galaxy instance. Currently no downloads or clean-up done. This is probably an error that you cannot recover automatically from.
5 Workflow scheduling failed at the Galaxy instance. Currently no downloads or clean-up done. This is probably an error that you cannot recover automatically from.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

galaxy-workflow-executor-0.2.6.tar.gz (13.3 kB view details)

Uploaded Source

Built Distribution

galaxy_workflow_executor-0.2.6-py3-none-any.whl (15.5 kB view details)

Uploaded Python 3

File details

Details for the file galaxy-workflow-executor-0.2.6.tar.gz.

File metadata

  • Download URL: galaxy-workflow-executor-0.2.6.tar.gz
  • Upload date:
  • Size: 13.3 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.4.2 importlib_metadata/4.6.3 pkginfo/1.7.1 requests/2.26.0 requests-toolbelt/0.9.1 tqdm/4.62.0 CPython/3.9.6

File hashes

Hashes for galaxy-workflow-executor-0.2.6.tar.gz
Algorithm Hash digest
SHA256 a9f2411a26bb42359bd6e6c2c4fe8e439f0504efa2ce52c460bc77af5f78a32b
MD5 40f82c2decd4edb3d09da683b17b80c4
BLAKE2b-256 15a300ae80e392f3f9ccd63c182f1b8baeba431916c66c6b63a49f868646a98b

See more details on using hashes here.

File details

Details for the file galaxy_workflow_executor-0.2.6-py3-none-any.whl.

File metadata

  • Download URL: galaxy_workflow_executor-0.2.6-py3-none-any.whl
  • Upload date:
  • Size: 15.5 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.4.2 importlib_metadata/4.6.3 pkginfo/1.7.1 requests/2.26.0 requests-toolbelt/0.9.1 tqdm/4.62.0 CPython/3.9.6

File hashes

Hashes for galaxy_workflow_executor-0.2.6-py3-none-any.whl
Algorithm Hash digest
SHA256 4ef5c78ca0c4c04ac1fa76fa1daf59fe66fed13de493e28e8b5098b910277033
MD5 86a01351f0c0ce2c9530c1e8c9d24cb3
BLAKE2b-256 6a4af4bacf3ede72a1dfea8946df6c970d643fdf8fa4f4c968166eaaa1d5554f

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page