Skip to main content

A python package for aligning and stitching light sheet fluorescence microscopy images

Project description

Rhapso

This is the official code base for Rhapso, a modular Python toolkit for the alignment and stitching of large-scale microscopy datasets.

Available on PyPI

License Python Version Documentation


Table of Contents



Update 1/12/26

Rhapso is still loading... and while we wrap up development, a couple things to know if you are outside the Allen Institute:

  • This process requires a very specific XML structure to work.
  • Fusion/Mutliscale is included but still under testing and development

Summary

Rhapso is a set of Python components used to register, align, and stitch large-scale, overlapping, tile-based, multiscale microscopy datasets. Its stateless components can run on a single machine or scale out across cloud-based clusters. Rhapso was developed by the Allen Institute for Neural Dynamics.


Contact

Questions or want to contribute? Please open an issue..


Supported Features

  • Interest Point Detection - DOG based feature detection
  • Interest Point Matching - Descriptor based RANSAC to match feature points
  • Global Optimization - Align matched features between tile pairs globally
  • Validation and Visualization Tools - Validate component specific results for the best output
  • ZARR - Zarr data as input
  • TIFF - TIFF data as input
  • AWS - AWS S3 based input/output and Ray based EC2 instances
  • Scale - Tested on 200 TB of data without downsampling


Layout

Rhapso/
└── Rhapso/
    ├── data_prep/                          # Custom data loaders
    ├── detection/
    ├── evaluation/
    ├── fusion/
    ├── image_split/
    ├── matching/
    ├── pipelines/
    │   └── ray/
    │       ├── aws/
    │       │   ├── config/                 # Cluster templates (edit for your account)
    │       │   └── alignment_pipeline.py   # AWS Ray pipeline entry point
    │       ├── local/
    │       │   └── alignment_pipeline.py   # Local Ray pipeline entry point
    │       ├── param/                      # Run parameter files (customize per run)
    │       ├── interest_point_detection.py # Detection pipeline script
    │       ├── interest_point_matching.py  # Matching pipeline script
    │       └── solver.py                   # Global solver script
    ├── solver/
    └── visualization/                      # Validation tools


Installation

Option 1: Install from PyPI (recommended)

# create and activate a virtual environment
python -m venv .venv && source .venv/bin/activate
# or: conda create -n rhapso python=3.10 && conda activate rhapso

# install Rhapso from PyPI
pip install Rhapso

Option 2: Install from GitHub (developers)

# clone the repo
git clone https://github.com/AllenNeuralDynamics/Rhapso.git

# create and activate a virtual environment
python -m venv .venv && source .venv/bin/activate
# or: conda create -n rhapso python=3.11 && conda activate rhapso

# install deps
pip install -r requirements.txt


How to Start

Rhapso is driven by pipeline scripts.

  • Each pipeline script has at minimum an associated param file (e.g. in Rhapso/pipelines/ray/param/).
  • If you are running on a cluster, you’ll also have a Ray cluster config (e.g. in Rhapso/pipelines/ray/aws/config/).

A good way to get started:

  1. Pick a template pipeline script
    For example:

    • Rhapso/pipelines/ray/local/alignment_pipeline.py (local)
    • Rhapso/pipelines/ray/aws/alignment_pipeline.py (AWS/Ray cluster)
  2. Point it to your param file
    Update the with open("...param.yml") line so it reads your own parameter YAML.

  3. (Optional) Point it to your cluster config
    If you’re using AWS/Ray, update the cluster config path.

  4. Edit the params to match your dataset
    Paths, downsampling, thresholds, matching/solver settings, etc.

  5. Run the pipeline
    The pipeline script will call the Rhapso components (detection, matching, solver, fusion) in the order defined in the script using the parameters you configured.



Try Rhapso on Sample Data

The quickest way to get familiar with Rhapso is to run it on a real dataset. We have a small (10GB) Z1 example hosted in a public S3 bucket, so you can access it without special permissions. It’s a good starting point to copy and adapt for your own alignment workflows.

XML (input)

  • s3://aind-open-data/HCR_802704_2025-08-30_02-00-00_processed_2025-10-01_21-09-24/image_tile_alignment/single_channel_xmls/channel_488.xml

Image prefix (referenced by the XML)

  • s3://aind-open-data/HCR_802704_2025-08-30_02-00-00_processed_2025-10-01_21-09-24/image_radial_correction/

Note: Occasionally we clean up our aind-open-data bucket. If you find this dataset does not exist, please create an issue and we will replace it.



High Level Approach to Registration, Alignment, and Fusion

This process has a lot of knobs and variations, and when used correctly, can work for a broad range of datasets.

First, figure out what type of alignment you need.

  • Are there translations to shift to?
  • If so, you’ll likely want to start with a rigid alignment.

Once you’ve run the rigid step, how does your data look?

  • Did the required translations shrink to an acceptable level?
  • If not, try again with new parameters, keeping the questions above in mind.

At this point, the translational part of your alignment should be in good shape. Now ask: are transformations needed? If so, you likely need an affine alignment next.

Your dataset should be correctly aligned at this point. If not, there are a number of reasons why, and we have listed some common recurrences and will keep this up to date.

There is a special case in some datasets where the z-stack is very large. In this case, you can use the split-dataset utility, which splits each tile into chunks. Then you can run split-affine alignment, allowing for more precise transformations without such imposing global rails.

Common Causes of Poor Alignment

  • Not enough quality matches (adjust sigma threshold until you do)
  • Data is not consistent looking (we take a global approach to params)
  • Large translations needed (extend search radius)
  • Translations that extend beyond overlapping span (increase overlap)


Performance

Interest Point Detection Performance Example (130TB Zarr dataset)

Environment Resources Avg runtime
Local single machine 10 CPU, 10 GB RAM ~120 min
AWS Ray cluster 560 CPU, 4.4 TB RAM ~30 min

*Actual times vary by pipeline components, dataset size, tiling, and parameter choices.*

Ray

Ray is a Python framework for parallel and distributed computing. It lets you run regular Python functions in parallel on a single machine or scale them out to a cluster (e.g., AWS) with minimal code changes. In Rhapso, we use Ray to process large scale datasets.

  • Convert a function into a distributed task with @ray.remote
  • Control scheduling with resource hints (CPUs, memory)

[!TIP] Ray schedules greedily by default and each task reserves 1 CPU, so if you fire many tasks, Ray will try to run as many as your machine advertises—often too much for a laptop. Throttle concurrency explicitly so you don’t overload your system. Use your machine's activity monitor to track this or the Ray dashboard to monitor this on your cluster:

  • Cap by CPUs:
    @ray.remote(num_cpus=3)   # Ray will schedule each time 3 cpus are available
    
  • Cap by Memory and CPU if Tasks are RAM-Heavy (bytes):
    @ray.remote(num_cpus=2, memory=4 * 1024**3)  # 4 GiB and 2 CPU per task>
    
  • No Cap on Resources:
    @ray.remote             
    
  • Good Local Default:
    @ray.remote(num_cpus=2)
    


Run Locally with Ray

1. Edit or create param file (templates in codebase)

Rhapso/Rhapso/pipelines/param/

2. Update alignment pipeline script to point to param file

with open("Rhapso/pipelines/ray/param/your_param_file.yml", "r") as file:
    config = yaml.safe_load(file)

3. Run local alignment pipeline script

python Rhapso/pipelines/ray/local/alignment_pipeline.py


Run on AWS Cluster with Ray

1. Edit/create param file (templates in codebase)

Rhapso/pipelines/ray/param/

2. Update alignment pipeline script to point to param file

with open("Rhapso/pipelines/ray/param/your_param_file.yml", "r") as file:
    config = yaml.safe_load(file)

3. Edit/create config file (templates in codebase)

Rhapso/pipelines/ray/aws/config/

5. Update alignment pipeline script to point to config file

unified_yml = "your_cluster_config_file_name.yml"

7. Run AWS alignment pipeline script

python Rhapso/pipelines/ray/aws/alignment_pipeline.py

[!TIP]

  • The pipeline script is set to always spin the cluster down, it is a good practice to double check in AWS.
  • If you experience a sticky cache on run params, you may have forgotten to spin your old cluster down.

Access Ray Dashboard

This is a great place to tune your cluster's performance.

  1. Find public IP of head node.
  2. Replace the ip address and PEM file location to ssh into head node.
    ssh -i /You/path/to/ssh/key.pem -L port:localhost:port ubuntu@public.ip.address
    
  3. Go to dashboard.
    http://localhost:8265
    


Parameters

Detection

| Parameter          | Feature / step         | What it does                                                                                  | Typical range\*                   |
| :----------------- | :--------------------- | :-------------------------------------------------------------------------------------------- | :-------------------------------- |
| `dsxy`             | Downsampling (XY)      | Reduces XY resolution before detection; speeds up & denoises, but raises minimum feature size | 16                                |
| `dsz`              | Downsampling (Z)       | Reduces Z resolution; often lower than XY due to anisotropy                                   | 16                                |
| `min_intensity`    | Normalization          | Lower bound for intensity normalization prior to DoG                                          | 1                                 |
| `max_intensity`    | Normalization          | Upper bound for intensity normalization prior to DoG                                          | 5                                 |
| `sigma`            | DoG blur               | Gaussian blur scale (sets feature size); higher = smoother, fewer peaks                       | 1.5 - 2.5                         |
| `threshold`        | Peak detection (DoG)   | Peak threshold (initial min peak ≈ `threshold / 3`); higher = fewer, stronger peaks           | 0.0008 - .05                      |
| `median_filter`    | Pre-filter (XY)        | Median filter size to suppress speckle/isolated noise before DoG                              | 1-10                              |
| `combine_distance` | Post-merge (DoG peaks) | Merge radius (voxels) to de-duplicate nearby detections                                       | 0.5                               |
| `chunks_per_bound` | Tiling/parallelism     | Sub-partitions per tile/bound; higher improves parallelism but adds overhead                  | 12-18                             |
| `max_spots`        | Post-cap               | Maximum detections per bound to prevent domination by dense regions                           | 8,0000 - 10,000                   |

Matching

# Candidate Selection
| Parameter                      | Feature / step      | What it does                                                      | Typical range  |
| :----------------------------- | :------------------ | :---------------------------------------------------------------- | :------------- |
| `num_neighbors`                | Candidate search    | Number of nearest neighbors to consider per point                 | 3              |
| `redundancy`                   | Candidate search    | Extra neighbors added for robustness beyond `num_neighbors`       | 0 - 1          |
| `significance`                 | Ratio test          | Strictness of descriptor ratio test; larger = stricter acceptance | 3              |
| `search_radius`                | Spatial gating      | Max spatial distance for candidate matches (in downsampled units) | 100 - 300      |
| `num_required_neighbors`       | Candidate filtering | Minimum neighbors required to keep a candidate point              | 3              |

# Ransac
| Parameter                     | Feature / step       | What it does                                                      | Typical range  |
| :---------------------------- | :------------------- | :---------------------------------------------------------------- | :------------- |
| `model_min_matches`           | RANSAC               | Minimum correspondences to estimate a rigid transform             | 18 – 32        |
| `inlier_threshold`            | RANSAC               | Inlier tolerance scaling; larger = looser inlier threshold        | 50 – 100       |
| `min_inlier_ratio`            | RANSAC               | Regularization strength during model fitting                      | 0.1 – 0.05     |
| `num_iterations`              | RANSAC               | Number of RANSAC trials; higher = more robust, slower             | 10,0000        |
| `regularization_weight`       | RANSAC               | Weight applied to the regularization term                         | 0.05 - 1.0     |


Solver

| Parameter                | Feature / step | What it does                                                       | Typical range       |
| :----------------------- | :------------- | :----------------------------------------------------------------- | :------------------ |
| `relative_threshold`     | Graph pruning  | Reject edges with residuals above dataset-relative cutoff          | 3.5                 |
| `absolute_threshold`     | Graph pruning  | Reject edges above an absolute error bound (detection-space units) | 7.0                 |
| `min_matches`            | Graph pruning  | Minimum matches required to retain an edge between tiles           | 3                   |
| `damp`                   | Optimization   | Damping for iterative solver; higher can stabilize tough cases     | 1.0                 |
| `max_iterations`         | Optimization   | Upper bound on solver iterations                                   | 10,0000             |
| `max_allowed_error`      | Optimization   | Overall error cap; `inf` disables hard stop by error               | `inf`               |
| `max_plateauwidth`       | Early stopping | Stagnation window before stopping on no improvement                | 200                 |
| `regularization_weight`  | RANSAC         | Weight applied to the regularization term                          | 0.05 - 1.0          |



Tuning Guide

  • Start with Detection. The quality and density of interest points strongly determine alignment outcomes.

  • Target Counts (exaSPIM): ~25–35k points per tile in dense regions; ~10k for sparser tiles. Going much higher usually increases runtime without meaningful accuracy gains.

  • Inspect Early. After detection, run the visualization script and verify that peaks form clustered shapes/lines with a good spatial spread—a good sign for robust rigid matches.

  • Rigid → Affine Dependency. Weak rigid matches produce poor rigid transforms, which then degrade affine matching (points don’t land close enough). If tiles fail to align:

    • Check match counts for the problem tile and its neighbors.
    • Adjust high-impact detection knobs—sigma, threshold, and median_filter—within sensible ranges.
    • Revisit max_spots and combine_distance to balance density vs. duplicate detections.


Build Package

Using the Built .whl File

  1. Build the .whl File in the root of this repo:
cd /path/to/Rhapso
pip install setuptools wheel
python setup.py sdist bdist_wheel

The .whl file will appear in the dist directory. Do not rename it to ensure compatibility (e.g., rhapso-0.1-py3-none-any.whl).





Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

rhapso-0.2.5.tar.gz (115.3 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

rhapso-0.2.5-py3-none-any.whl (134.0 kB view details)

Uploaded Python 3

File details

Details for the file rhapso-0.2.5.tar.gz.

File metadata

  • Download URL: rhapso-0.2.5.tar.gz
  • Upload date:
  • Size: 115.3 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for rhapso-0.2.5.tar.gz
Algorithm Hash digest
SHA256 74d2efaf3692dade8a64142001ca499b0ce3b23cd2909aeb739435523d5b5fe7
MD5 73cb1ba8f98819c381c8a8f502076671
BLAKE2b-256 e9873e0e1c389c3b279aa55eaab9cef92f665c0ef2cfb3695edbd2255ba822b9

See more details on using hashes here.

File details

Details for the file rhapso-0.2.5-py3-none-any.whl.

File metadata

  • Download URL: rhapso-0.2.5-py3-none-any.whl
  • Upload date:
  • Size: 134.0 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for rhapso-0.2.5-py3-none-any.whl
Algorithm Hash digest
SHA256 13847ccc3c7bfcf52f7bd6cdd415dd05fbe68494e4cdd013445e07b0fe4e80dc
MD5 44ca5b7881f2293d02b67f7fc94291b9
BLAKE2b-256 e1a9d8198b36ed577442b827f6a2c66bce66b411b499eade66e08ec8d8a3c150

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page