Skip to main content

Services for CryoEM processing

Project description

cryoem-services

PyPI release Supported Python versions Test coverage Code style: black

Services and configuration for cryo-EM pipelines.

This package consists of a number of services to process cryo-EM micrographs, both for single particle analysis and tomography, using a range of commonly used cryo-EM processing software. These services can be run independently to process data, or as part of a wider structure for performing live analysis during microscope collection. For live analysis, this package integrates with a package for transferring and monitoring collected data, Murfey, and a database for storing processing outcomes, ISPyB.

To run these services the software executables being called must be installed. These do not come with this package.

Tomography processing

The tomography processing pipeline consists of:

  • Motion correction
  • CTF estimation
  • Tomogram alignment
  • Tomogram denoising using Topaz
  • Segmentation using membrain-seg

The results of this processing can be opened and continued using Relion 5.0.

Single particle analysis

The single particle analysis pipeline produces a project that can be opened and continued using CCP-EM doppio or Relion.

The processing pipeline consists of:

  • Motion correction
  • CTF estimation
  • Particle picking
  • (Optionally) Ice thickness estimation
  • Particle extraction and rebatching
  • 2D classification using Relion
  • Automated 2D class selection using Relion
  • 3D classification using Relion
  • 3D Refinement and post-processing
  • BFactor estimation by refinement with varying particle count

Services currently available

The following services are provided for running the pipelines:

  • Utility services:
    • ClusterSubmission: Submits wrapper scripts to an HPC cluster
    • EMISPyB: Inserts results into an ISPyB database
    • Images: Creates thumbnail images for viewing processing outcomes
    • NodeCreator: Creates Relion project files for the services run
    • ProcessRecipe: Converts recipes into messages suitable for processing services
  • Processing services:
    • BFactor: Performs the setup for 3D refinement with varying particle count
    • CrYOLO: Particle picking on micrographs using crYOLO
    • CTFFind: CTF estimation on micrographs using CTFFIND4
    • DenoiseSlurm: Tomogram denoising, submitted to a slurm HPC cluster, using Topaz
    • Extract: Extracts picked particles from micrographs
    • ExtractClass: Extracts particles from a given 3D class
    • IceBreaker: Ice thickness estimation with IceBreaker
    • MembrainSeg: Tomogram segmentation, submitted to a slurm HPC cluster, using membrain-seg
    • MotionCorr: Motion correction of micrographs using MotionCor2 or Relion, optionally submitted to a slurm HPC cluster
    • PostProcess: Post-processing of 3D refinements using Relion
    • SelectClasses: Runs automated 2D class selection using Relion and re-batches the particles from these classes
    • SelectParticles: Creates files listing batches of extracted particles
    • TomoAlign: Tomogram reconstruction from a list of micrographs using imod and AreTomo2
    • TomoAlignSlurm: Tomogram alignment processing submitted to a slurm HPC cluster

There are also three wrapper scripts that can be run on an HPC cluster using the ClusterSubmission service. These perform 2D classification, 3D classification and 3D refinement using Relion.

Running services

The services in this package are run using python-workflows. They consume messages off a RabbitMQ instance and processing happens in sequences defined by the recipes in the recipes folder. To start a service run the cryoemservices.service command and specify the service name. For example, to start a motion correction service:

$ cryoemservices.service -s MotionCorr -c config_file.yaml

The configuration file should contain the following:

rabbitmq_credentials: <file with connection credentials to rabbitmq>
recipe_directory: <directory containing the recipes to run>
ispyb_credentials: <(Optionally) file with credentials for an ispyb database>
slurm_credentials:
  default: <(Optionally) credentials for a slurm RestAPI>
graylog_host: <(Optionally) the name of a graylog instance>
graylog_port: <(Optionally) the port used by graylog>

Once started, these services will initialise and then wait for messages to be sent to them. Messages are sent through a message broker, currently RabbitMQ is supported using pika transport in python-workflows. Individual processing stages can be run by sending a dictionary of the parameters, but the processing pipelines are designed to run through recipes.

A recipe is a specication of a series of steps to carry out, and how these steps interact with each other. Recipes for the current processing pipelines are provided in the recipes folder.

To run a recipe in python a dictionary needs to be provided consisting of the recipe name and the parameters expected by the recipe. The following snippet shows an example of the setup needed. This will send a message to a running ProcessRecipe service which prepares the recipe for the processing services.

import workflows.transport.pika_transport as pt

example_message = {
    "recipes": ["em-tomo-align"],
    "parameters": {
        "path_pattern": "micrograph_*.mrc",
        "pix_size": "1",
        ...
    },
}

transport = pt.PikaTransport()
transport.connect()
transport.send("processing_recipe", example_message)

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

cryoemservices-1.2.0.tar.gz (129.5 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

cryoemservices-1.2.0-py3-none-any.whl (167.7 kB view details)

Uploaded Python 3

File details

Details for the file cryoemservices-1.2.0.tar.gz.

File metadata

  • Download URL: cryoemservices-1.2.0.tar.gz
  • Upload date:
  • Size: 129.5 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.12.9

File hashes

Hashes for cryoemservices-1.2.0.tar.gz
Algorithm Hash digest
SHA256 d28d5f6658cff54e8b67c95b706203d58fc1928569f89146f378d8510a85f798
MD5 6fee02e8a9bb774e4ced03f6983b5d88
BLAKE2b-256 8b53a7d1cf0c001b22a92e26eedf0ccdddd4cfe830de8fd8805c1c8d14a962c9

See more details on using hashes here.

Provenance

The following attestation bundles were made for cryoemservices-1.2.0.tar.gz:

Publisher: publish-version.yml on DiamondLightSource/cryoem-services

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file cryoemservices-1.2.0-py3-none-any.whl.

File metadata

  • Download URL: cryoemservices-1.2.0-py3-none-any.whl
  • Upload date:
  • Size: 167.7 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.12.9

File hashes

Hashes for cryoemservices-1.2.0-py3-none-any.whl
Algorithm Hash digest
SHA256 bd5206010fed94e861ff3c9a6bdf241c74e1885cf3757e6375aa3ff47be4dd22
MD5 ceb6bab8dfda813ef07c567db5aeb47e
BLAKE2b-256 b6889f57f844997f70188fde6e12a85e9619ac0abe8e942c93e72b0d708bcb67

See more details on using hashes here.

Provenance

The following attestation bundles were made for cryoemservices-1.2.0-py3-none-any.whl:

Publisher: publish-version.yml on DiamondLightSource/cryoem-services

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page