Skip to main content

A framework to easily train deep learning model on Imaging Atmospheric Cherenkov Telescopes data

Project description

GammaLearn

Deep Learning for Imaging Cherenkov Telescopes Data Analysis.

GammaLearn is a collaborative project to apply deep learning to the analysis of low-level Imaging Atmospheric Cherenkov Telescopes such as CTA. It provides a framework to easily train and apply models from a configuration file.

DOI

pipeline status coverage report

Table of Contents

  1. Installation
  2. For users
  3. For Developers
  4. Contributing
  5. Cite Us
  6. License

Usage

Gammalearn's CI/CD pipelines produce containers that can be used to quickly get a working environment for development or production. The containers are uploaded to the gammalearn's container registry. A set of containers is available for each branch in the repository, and the containers are re-build for any commit that involves modifications made to the environment or dockerfiles files (anything in the docker directory).

Gammalearn containers include the pytorch and cuda dependencies to do computation on nvidia gpus, in addition to a big stack of scientific computing software. They are typically take several GBs.

For users

We recommend the use of apptainer. To get the production image of the version of gammalearn you want to use, for instance to get gammalearn v0.13.0, use apptainer pull:

apptainer pull docker://gitlab-registry.in2p3.fr/gammalearn/gammalearn:v0.13.0

This will create a .sif container file that contains a ready to use gammalearn installation. Warning: apptainer can use several GB of disk space as cache when building the .sif file. By default, the cache is located in your home folder ~/.apptainer/cache. You can change this location by setting the APPTAINER_CACHEDIR environment variable. Clean apptainer's cache with apptainer cache clean

You can now get a shell inside the container to test gammalearn:

apptainer shell path_to_your_sif_file.sif
Apptainer> /opt/conda/bin/gammalearn --help

Run an experiment

You can run an experiment using apptainer run. Since apptainer containers are read-only by default, you will need to mount the paths to your input and output files. To use nvidia gpus, you will need to specify the --nv option as well. A typical command example:

# Run the experiment in the container
# Parameters:
# --nv                  to use nvidia gpus from inside the container
# CUDA_VISIBLE_DEVICES  env variable used by pytorch to discover the gpus
# NUMBA_CACHE_DIR       a writable directory where numba can store its compiled functions
#                       (needs to be outside of the container, which is read-only)
# CTAPIPE_CACHE         ctapipe needs a writable place, to store its downloaded files.
# Mounts: input (data and settings file) and output directories
# 
# We call the gammalearn entrypoint directly in /opt/conda/bin, because micromamba is not initialized
# inside the container for a new user (and every user is new, since with apptainer the user remains the same
# as the user on the host system by default (only users defined in the containers are known))
apptainer run \
    --nv \
    --env "CUDA_VISIBLE_DEVICES=$CUDA_VISIBLE_DEVICES" \
    --env "NUMBA_CACHE_DIR=/tmp/NUMBA" \
    --env "CTAPIPE_CACHE=/tmp/CTAPIPE" \
    --mount type=bind,source=/path/to/input/data_dir/,destination=/corresponding/path/in/container/ \
	--mount type=bind,source=/path/to/output/data_dir/,destination=/corresponding/path/in/container/ \
    path_to_your_sif_file.sif /opt/conda/bin/gammalearn path_to_your_experiment_settings_file.py

You can find examples of setting file in the examples and some sample data in example data

For Developers

Developers are encouraged to use the development containers available in the repository container registry. The development containers contain all gammalearn run-time dependencies, and several tools required to develop the software (for instance to run tests, build the documentation, visualize results, etc.), but it does NOT contain gammalearn nor it source code. The goal is to provide a containerized development environment, but leave the source code in the host system. It is recommended to use the development containers with docker, either directly or through an IDE.

Development container with docker directly

To pull the container of the branch you use to develop:

docker pull gitlab-registry.in2p3.fr/gammalearn/gammalearn/dev:your_branch_name

Clone gammalearn locally and go in the gammalearn directory:

git clone https://gitlab.in2p3.fr/gammalearn/gammalearn.git && cd gammalearn

Enter the container while mounting the source code, and install gammalearn

# - get a shell with interactive mode
# - mount gammalearn sources in "bind" mode with --mount
# - use your host user uid and gid (-u $(id -u):$(id -g)) to have write permissions in the mounted gammalearn sources
# - set working directory to /src
# - optionally: set pull policy to always to always get the latest image for your branch
docker run --rm -it --mount type=bind,source=.,destination=/src -w /src --pull=always -u $(id -u):$(id -g) gitlab-registry.in2p3.fr/gammalearn/gammalearn/dev:your_branch_name
(base) mambauser@...:/src$ pip install --no-deps -e .
# Develop !

Usage with an IDE (vscode example)

Many IDE's offer the possibility to start or interact with running docker containers. In VScode, this is handled by the "dev container" extension included in the remote development extension pack (see the extension documentation). The extension can start a container and automatically install a vs-code server inside the container, allowing to transparently develop the software while using the environment from inside the container. The following devcontainer.json configuration allows to start a development container with the right permissions and build gammalearn from the mounted sources. As examples, the python extension and ruff linter are installed in the vs-code server running in the container.

// For format details, see https://aka.ms/devcontainer.json. For config options, see the
// README at: https://github.com/devcontainers/templates/tree/main/src/miniconda
{
	"image": "gitlab-registry.in2p3.fr/gammalearn/gammalearn/dev:your_dev_branch",
	"workspaceMount": "source=${localWorkspaceFolder},target=/src,type=bind",
	"workspaceFolder": "/src",
	"postAttachCommand": "micromamba run -n base pip install --no-deps -e /src",
	"runArgs": [ "--network=host"],
	"customizations": {
		// Configure properties specific to VS Code.
		"vscode": {
			// Add the IDs of extensions you want installed when the container is created.
			"extensions": [
				"charliermarsh.ruff",
				"ms-python.python",
				"ms-python.vscode-pylance",
				"njpwerner.autodocstring",
				"tamasfe.even-better-toml",
				"wmaurer.change-case"
			]
		}
	}
}

Note: By default, the dev containers extension will not re-pull the specified image unless you rebuild the container in vscode. Make sure to pull the latest container for your branch before entering the container.

Once the container is running, you can enter it from another external terminal with

# Get the container ID of your dev container started with vscode
docker ps
# Get a shell in the container
docker exec -it -w /src container_ID bash​

Bare-metal installation

Gammalearn dependencies are managed using Anaconda. The dependencies are split into run-time, test and development dependencies available in the docker directory. According to your usage, install the dependencies you need. For instance:

conda env create -n glearn -f dependencies.yml
conda env update -n glearn -f test_dependencies.yml
conda activate glearn

Note for OSX and/or no-gpu users: please edit the environment file to remove cudatoolkit from the dependencies.

Older version installation

Create the gammalearn release environment:

VERSION=0.12.0
wget https://gitlab.in2p3.fr/gammalearn/gammalearn/-/raw/v${VERSION}/environment.yml -O glearn_${VERSION}_env.yml
conda env create -f glearn_${VERSION}_env.yml
conda activate glearn

Note for OSX and/or no-gpu users: please edit the environment file to remove cudatoolkit from the dependencies.

Contributing

Contributions are very much welcome: please see CONTRIBUTING.

Cite Us

Please cite

Jacquemont M, Vuillaume T, Benoit A, Maurin G, Lambert P, Lamanna G, Brill A. GammaLearn: A Deep Learning Framework for IACT Data. In36th International Cosmic Ray Conference (ICRC2019) 2019 Jul (Vol. 36, p. 705). DOI: https://doi.org/10.22323/1.358.0705

For reproducibility purposes, please also cite the exact version of GammaLearn you used by citing the corresponding DOI on Zenodo:
DOI

License

GammaLearn is distributed under an MIT license.

## Back to top

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

gammalearn-0.15.0.tar.gz (2.0 MB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

gammalearn-0.15.0-py3-none-any.whl (435.9 kB view details)

Uploaded Python 3

File details

Details for the file gammalearn-0.15.0.tar.gz.

File metadata

  • Download URL: gammalearn-0.15.0.tar.gz
  • Upload date:
  • Size: 2.0 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.11.13

File hashes

Hashes for gammalearn-0.15.0.tar.gz
Algorithm Hash digest
SHA256 2ff2893ae247f5618ac2db75ca3e5233b748f336eb6270f524e19a5af5502091
MD5 df89396d9c303dd7c7f0046d5924edf8
BLAKE2b-256 33a9e3d4095315e4c01e502f9c4ec11c653652e71fafb17fc673163697dcabdc

See more details on using hashes here.

File details

Details for the file gammalearn-0.15.0-py3-none-any.whl.

File metadata

  • Download URL: gammalearn-0.15.0-py3-none-any.whl
  • Upload date:
  • Size: 435.9 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.11.13

File hashes

Hashes for gammalearn-0.15.0-py3-none-any.whl
Algorithm Hash digest
SHA256 3d67515fb002f24ce7c85c46dd93bf305c8794579075c45e732627d49945d427
MD5 033be375f842ffc9d23017a1b78a0d8b
BLAKE2b-256 69fb02950f54b464eb02fa2311941f60add352ee51ffb66e88e4f69a71e00fce

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page