Skip to main content

CLI and utils for the Neuro-iX lab

Project description

Neuro-iX Tools

Common Tools for the Neuro-iX Lab

Getting Started

Installation

You will need an environment with at least Python 3.11, then run:

pip install neuro-ix-tools

Alternatively, you can clone the repository and use:

python cli.py

instead of neuro-ix.

Setup

If you are using the package, simply run:

neuro-ix init

This provides sensible defaults for Narval. The configuration file is stored in your .config folder.

If you are using the repository directly, we recommend using a local .env file. A template is available at .example.env.

Usage

Inside this environment, you have access to the neuro-ix command, which currently exposes one main tool.

FreeSurfer recon-all on SLURM Cluster

We provide a pipeline that simplifies the usage of FreeSurfer on the Narval SLURM cluster. The main command is:

neuro-ix freesurfer recon-all

This allows users to process all subjects in either a BIDS or CAPS (Clinica) dataset with FreeSurfer, using one SLURM job per subject.

Arguments:
  • --bids-dataset: Path to the root of a BIDS-compliant dataset
  • --clinica-dataset: Path to the root of a Clinica-compliant dataset (CAPS)
  • --cortical-stats: Flag to store only FreeSurfer's stats files
  • --start-from: Used when there are more than 1000 subjects due to Narval’s job limit. Allows the user to resume processing from a specific subject index.
Example:
neuro-ix freesurfer recon-all --bids-dataset /path/to/dataset

If your dataset includes more than 1000 subjects (e.g., 1500), once the first batch is done, run:

neuro-ix freesurfer recon-all --bids-dataset /path/to/dataset --start-from 1000

Library

As a library, the neuro_ix package exposes:

  • Classes to interact with and query BIDS and CAPS datasets for T1-weighted MRIs
  • Extendable command classes

Contributing

Setup

Once the repository is cloned, install the development dependencies with:

pip install -r dev_requirements.txt

Tests

Test Tools

We use:

  • pytest for unit tests
  • pytest-cov for coverage reports
    Run tests via:
pytest --cov
  • ruff for linting and formatting (automatically applied via pre-commit)
  • Additional tools for code quality: ssort, pydocstyle, mypy, and pylint

Test Data

All test data are extracted from MR-ART:

Nárai, Á., Hermann, P., Auer, T. et al. Movement-related artefacts (MR-ART) dataset of matched motion-corrupted and clean structural MRI brain scans. Sci Data 9, 630 (2022). https://doi.org/10.1038/s41597-022-01694-8

Deployment

Build Package using :

python -m build

And deploy to PyPI with :

twine upload sit/*

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

neuro_ix_tools-0.0.8.tar.gz (11.8 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

neuro_ix_tools-0.0.8-py3-none-any.whl (11.3 kB view details)

Uploaded Python 3

File details

Details for the file neuro_ix_tools-0.0.8.tar.gz.

File metadata

  • Download URL: neuro_ix_tools-0.0.8.tar.gz
  • Upload date:
  • Size: 11.8 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.11.11

File hashes

Hashes for neuro_ix_tools-0.0.8.tar.gz
Algorithm Hash digest
SHA256 1bfa8a0268d3eda2922f7392410326dcc91218ed0e31015a36b015921e77dedd
MD5 0a3e0b580e07dba83339caf6cc7ba632
BLAKE2b-256 4917573a691112e54f9f167658fd93a593c4e367a36b311f8918ecbe54c8da2e

See more details on using hashes here.

File details

Details for the file neuro_ix_tools-0.0.8-py3-none-any.whl.

File metadata

  • Download URL: neuro_ix_tools-0.0.8-py3-none-any.whl
  • Upload date:
  • Size: 11.3 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.11.11

File hashes

Hashes for neuro_ix_tools-0.0.8-py3-none-any.whl
Algorithm Hash digest
SHA256 c33e0e5b4049d5c712ee1475de9cd88c76f25f51e18e79855daa5850d00cc617
MD5 cd22d5e52b39bd17fa37d0204c11c138
BLAKE2b-256 2d2f1ef49df9886fccc20253f57f2fb61e45b0ab64ac2b0fcaa5b39fb6a5a22c

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page