Skip to main content

Deep Learning based tool to quantify subject motion in T1w brain MRI

Project description

Agitation

This repository presents a deep learning-based tool to quantify subject motion in T1-weighted brain MRI.
The model used by this tool can be trained using our research code.

Getting Started

Installation

You will need an environment with at least Python 3.11. Then run:

pip install agitation

Alternatively, you can clone the repository and use:

python cli.py

instead of agitation.

Setup

Model

We use a TorchScript version of our best model.
All model checkpoints and the final TorchScript file are available on Zenodo.

The model will be downloaded automatically when needed. However, you can also manually download it with:

agitation manage check

The model is stored in your application data directory. You can retrieve the exact location using the check command.

To remove all downloaded data:

agitation manage delete

MRI Data

Our model was trained on data preprocessed with Clinica's T1-linear pipeline.
While it may work with any T1-weighted MRI, we strongly recommend using the same preprocessing pipeline to ensure consistent results.

CLI Usage

Full Dataset

To quantify motion on a full dataset, use the command:

agitation dataset
Arguments:
  • -d, --dataset: Path to the root of the dataset. It must be organized according to BIDS or CAPS (Clinica) standards and contain either an anat folder or t1_linear for CAPS.
  • -f, --file: Path to a CSV file describing the data to process. The file must contain at least a data column specifying the path to each volume. Other columns will be copied to the output CSV.
  • -g, --gpu: Flag to enable GPU inference.
  • --cuda: Specify the GPU index to use (defaults to 0).
  • -o, --output: Path to the output CSV file.
Examples:
agitation dataset --dataset <path_to_root> -g --output <path_to_output_file>
agitation dataset --file <path_to_csv>

Subject Level

To quantify motion at the subject level, use the command:

agitation inference

This entry point is recommended for use with pipeline tools like Nipoppy.

Arguments:
  • --bids_dir: Path to the root of a BIDS dataset.
  • --subject_id: Subject identifier in BIDS format: sub-<label>.
  • --session_id: Session identifier in BIDS format: ses-<label>.
  • -g, --gpu: Flag to enable GPU inference.
  • --cuda: Specify the GPU index to use (defaults to 0).
  • --output_dir: Path to the output directory.
Example:
agitation inference --bids_dir tests/data/bids_sub_ses --subject_id sub-000103 --session_id ses-standard --output_dir ./

Container, Boutiques, and Nipoppy

Our tool offers a Boutiques descriptor, available in the descriptors folder, and an Apptainer container (container definition in containers).

To integrate our tool into a Nipoppy dataset, copy descriptors/agitation.json to your dataset's pipelines/agitation-<version>/descriptor.json and descriptors/invocation.json to pipelines/agitation-<version>/invocation.json.

Library

The agitation package can also be used as a library to include motion estimation in your projects.

Downloading the Model

To manually download the model within your code:

from agitation.data_manager import download_model

download_model()

Dataloader Inference

To run inference on a dataloader:

from monai.data.dataset import Dataset
from torch.utils.data import DataLoader

from agitation.inference import estimate_motion_dl
from agitation.processing import LoadVolume

# Example usage
dataset = Dataset(<your_data_as_a_dict>, transform=LoadVolume())
dataloader = DataLoader(dataset)
estimate_motion_dl(dataloader, cuda=0)

Batch Inference

To perform inference on a single batch:

import torch

from agitation.config import MODEL_PATH
from agitation.processing import SoftLabelToPred

# Dataloading, cropping, and normalization steps

model = torch.jit.load(
    MODEL_PATH,
    map_location="cuda:0"  # If using CUDA
)
converter = SoftLabelToPred()

with torch.inference_mode():
    prediction = model(data).cpu()
    motions = converter(prediction)

Contributing

Setup

Once the repository is cloned, install the development dependencies with:

pip install -r dev_requirements.txt

Tests

Test Tools

We use:

  • pytest for unit tests
  • pytest-cov for coverage reports (targeting 100% test coverage)

Run tests via:

pytest --cov

Other tools:

  • ruff for linting and formatting (automatically applied via pre-commit)
  • Additional code quality tools: ssort, pydocstyle, mypy, and pylint

Test Data

All test data are extracted from MR-ART:

Nárai, Á., Hermann, P., Auer, T. et al. Movement-related artefacts (MR-ART) dataset of matched motion-corrupted and clean structural MRI brain scans. Sci Data 9, 630 (2022). https://doi.org/10.1038/s41597-022-01694-8

Deployment

To fully deploy a new version, follow these steps in order:

  1. Build and deploy the PyPI package (used for the Apptainer image).
  2. Build the Apptainer image and publish it to Docker Hub.
  3. Publish any modifications to the Boutiques descriptor on Zenodo.

Python Packaging

Build the package using:

python -m build

Deploy to PyPI with:

twine upload dist/*

Apptainer Container

Build the container using:

apptainer build agitation.sif containers/agitation.def

Publish with:

apptainer push agitation.sif oras://docker.io/chbricout/agitation:latest

Boutiques Descriptor

Publish to Zenodo using:

bosh publish --sandbox -y --zenodo-token <ZENODO TOKEN> descriptors/agitation.json

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

agitation-0.0.2.tar.gz (26.5 MB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

agitation-0.0.2-py3-none-any.whl (11.2 kB view details)

Uploaded Python 3

File details

Details for the file agitation-0.0.2.tar.gz.

File metadata

  • Download URL: agitation-0.0.2.tar.gz
  • Upload date:
  • Size: 26.5 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.11.11

File hashes

Hashes for agitation-0.0.2.tar.gz
Algorithm Hash digest
SHA256 950bcdf4393cdc1f033d9577b462769b8c9ee00c02e1827cf1ceac73a475ece9
MD5 cbbf218237ae5587c2f77929723f69e9
BLAKE2b-256 8284a0bfb7a3150a4f18b5c4346e3e73b80dbded040ddbfb26fe75357023d7b1

See more details on using hashes here.

File details

Details for the file agitation-0.0.2-py3-none-any.whl.

File metadata

  • Download URL: agitation-0.0.2-py3-none-any.whl
  • Upload date:
  • Size: 11.2 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.11.11

File hashes

Hashes for agitation-0.0.2-py3-none-any.whl
Algorithm Hash digest
SHA256 63f37789a91589fae0d6895208229a92e0bfddf9debe18c7c018f199d0bd5cf8
MD5 766bb30569f80b6702a9240b6af217cf
BLAKE2b-256 4340913fb543f2114e38fde25b7ad5d80122ab850e964994f160959f0898de15

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page