Skip to main content

Python library for cloud and cloud shadow segmentation in high to moderate resolution satellite imagery

Project description

OmniCloudMask

image image image Conda Downloads image Tutorials

OmniCloudMask is a Python library for state-of-the-art cloud and cloud shadow segmentation in high to moderate resolution satellite imagery.

As a successor to the CloudS2Mask library, OmniCloudMask offers higher accuracy while supporting a wide range of resolutions, sensors, and processing levels.

OmniCloudMask has been validated on Sentinel-2, PlanetScope and Landsat data and is also known to work well with Maxar data, it should work on any imagery with Red Green and NIR bands with a spatial resolution of 50 m or better.

OmniCloudMask paper 📜

OmniCloudMask training data distribution map 🗺️

Satellite Image Deep Learning podcast about OmniCloudMask 🎙️

Changelog

See the changelog for version history and release notes.

See the model changelog for model version notes.

Features

  • Process imagery resolutions from 10 m to 50 m, (higher resolutions can be down sampled to 10 m).
  • Any imagery processing level.
  • Patch-based processing of large satellite images.
  • Multi-threaded patch compilation and model inference.
  • Option to export confidence maps.
  • Only requires Red, Green and NIR bands.
  • Known to work well with Sentinel-2, Landsat 8, PlanetScope and Maxar.
  • Supports inference on cuda, mps and cpu.
  • Model compilation for faster inference.

Try in Colab

Colab_Button

Example notebooks

How it works

Sensor agnostic Deep Learning with OmniCloudMask

Installation

To install the package, use one of the following command:

uv add omnicloudmask
pip install omnicloudmask
conda install conda-forge::omnicloudmask
pip install git+https://github.com/DPIRD-DMA/OmniCloudMask.git

Docker

Alternatively you can install OmniCloudMask within a Docker container by following the Docker instructions

Usage

Predict from array

To predict cloud and cloud shadow masks from a numpy array representing the Red, Green, and NIR bands, predictions are returned as a numpy array:

import numpy as np
from omnicloudmask import predict_from_array

# Example input array, in practice this should be Red, Green and NIR bands
input_array = np.random.rand(3, 1024, 1024)

# Predict cloud and cloud shadow masks
pred_mask = predict_from_array(input_array)

Predict from load function

To predict cloud and cloud shadow masks for a list of Sentinel-2 scenes, predictions are saved to disk along side the inputs as geotiffs, a list of prediction file paths is returned:

Sentinel-2

from pathlib import Path
from omnicloudmask import predict_from_load_func, load_s2

# Paths to scenes (L1C and or L2A)
scene_paths = [Path("path/to/scene1.SAFE"), Path("path/to/scene2.SAFE")]

# Predict masks for scenes
pred_paths = predict_from_load_func(scene_paths, load_s2)

Landsat

from pathlib import Path
from omnicloudmask import predict_from_load_func, load_ls8

# Paths to scenes
scene_paths = [Path("path/to/scene1"), Path("path/to/scene2")]

# Predict masks for scenes
pred_paths = predict_from_load_func(scene_paths, load_ls8)

Seep optimised options (for GPU)

pred_paths = predict_from_load_func(scene_paths=scene_paths, 
                                    load_func=load_s2,
                                    inference_dtype='bf16',
                                    compile_models=True,
                                    batch_size=4)

Low VRAM options

import torch
# Set this to the number of CPU cores if using mosaic_device='cpu'
torch.set_num_threads(4)

pred_paths = predict_from_load_func(scene_paths=scene_paths,
                                    load_func=load_s2,
                                    inference_dtype='bf16',
                                    batch_size=1,
                                    mosaic_device='cpu')

CPU inference

pred_paths = predict_from_load_func(scene_paths=scene_paths,
                                    load_func=load_s2,
                                    inference_dtype='fp32', # this is important for CPU inference
                                    batch_size=1,
                                    inference_device='cpu',
                                    mosaic_device='cpu')

Output

  • Output classes are defined by the CloudSEN12 paper and dataset used for training.
  • 0 = Clear
  • 1 = Thick Cloud
  • 2 = Thin Cloud
  • 3 = Cloud Shadow

Usage tips

  • If using an NVIDIA GPU make sure to increase the default 'batch_size'.
  • In most cases setting 'inference_dtype' to "bf16" should improve processing speed, if your hardware supports it.
  • If you are running out of VRAM even with a batch_size of 1 try setting the 'mosaic_device' device to 'cpu'.
  • Make sure if you are using imagery above 10 m res to downsample it before passing it to OmniCloudMask.
  • If you are processing many files try to use the 'predict_from_load_func' as it preloads data during inference, resulting in faster processing.
  • In some rare cases OmniCloudMask may fail to detect cloud if the raster data is clipped by sensor saturation or preprocessing, this results in image regions with no remaining texture to enable detection. To resolve this simply preprocess these regions and set the areas to 0, the no data value.
  • OmniCloudMask expects Red, Green and NIR bands, however if you don't have a NIR band then we have seen reasonable results passing Red Green BLUE bands into the model instead.
  • If you are processing more than 10-20 scenes using predict_from_load_func try turning on 'compile_models' it should reduce processing times by 10-20%.

Parameters

predict_from_load_func

  • scene_paths (Union[list[Path], list[str]]): A list of paths to the scene files to be processed.
  • load_func (Callable): A function to load the scene data.
  • patch_size (int): Size of the patches for inference. Defaults to 1000.
  • patch_overlap (int): Overlap between patches for inference. Defaults to 300.
  • batch_size (int): Number of patches to process in a batch. Defaults to 1.
  • inference_device (Union[str, torch.device]): Device to use for inference (e.g., 'cpu', 'cuda'). Defaults to None then default_device().
  • mosaic_device (Union[str, torch.device]): Device to use for mosaicking patches. Defaults to None then default_device().
  • inference_dtype (Union[torch.dtype, str]): Data type for inference. Defaults to torch.float32.
  • export_confidence (bool): If True, exports confidence maps instead of predicted classes. Defaults to False.
  • softmax_output (bool): If True, applies a softmax to the output, only used if export_confidence = True. Defaults to True.
  • no_data_value (int): Value within input scenes that specifies no data region. Defaults to 0.
  • overwrite (bool): If False, skips scenes that already have a prediction file. Defaults to True.
  • apply_no_data_mask (bool): If True, applies a no-data mask to the predictions. Defaults to True.
  • output_dir (Optional[Union[Path, str]], optional): Directory to save the prediction files. Defaults to None. If None, the predictions will be saved in the same directory as the input scene.
  • custom_models (Union[list[torch.nn.Module], torch.nn.Module], optional): A list or singular custom torch models to use for prediction. Defaults to None.
  • pred_classes (int, optional): Number of classes to predict. Defaults to 4, to be used with custom models. Defaults to 4.
  • destination_model_dir (Union[str, Path, None]): Directory to save the model weights. Defaults to None.
  • model_download_source (str, optional): Source from which to download the model weights. Defaults to "hugging_face", can also be "google_drive".
  • compile_models (bool, optional): If True, compiles the models for faster inference. Defaults to False.
  • compile_mode (str, optional): Compilation mode for the models. Defaults to "default".
  • model_version (float, optional: Version of the model to use. Defaults to the latest available version. Can also be set to 4.0, 3.0, 2.0, or 1.0 for older models.

predict_from_array

  • input_array (np.ndarray): A numpy array with shape (3, height, width) representing the Red, Green, and NIR bands.
  • patch_size (int): Size of the patches for inference. Defaults to 1000.
  • patch_overlap (int): Overlap between patches for inference. Defaults to 300.
  • batch_size (int): Number of patches to process in a batch. Defaults to 1.
  • inference_device (Union[str, torch.device]): Device to use for inference (e.g., 'cpu', 'cuda'). Defaults to None then default_device().
  • mosaic_device (Union[str, torch.device]): Device to use for mosaicking patches. Defaults to None then default_device().
  • inference_dtype (Union[torch.dtype, str]): Data type for inference. Defaults to torch.float32.
  • export_confidence (bool): If True, exports confidence maps instead of predicted classes. Defaults to False.
  • softmax_output (bool): If True, applies a softmax to the output, only used if export_confidence = True. Defaults to True.
  • no_data_value (int): Value within input scenes that specifies no data region. Defaults to 0.
  • apply_no_data_mask (bool): If True, applies a no-data mask to the predictions. Defaults to True.
  • custom_models (Union[list[torch.nn.Module], torch.nn.Module], optional): A list or singular custom torch models to use for prediction. Defaults to None.
  • pred_classes (int, optional): Number of classes to predict. Defaults to 4, to be used with custom models. Defaults to 4.
  • destination_model_dir (Union[str, Path, None]) : Directory to save the model weights. Defaults to None.
  • model_download_source (str, optional): Source from which to download the model weights. Defaults to "hugging_face", can also be "google_drive".
  • compile_models (bool, optional): If True, compiles the models for faster inference. Defaults to False.
  • compile_mode (str, optional): Compilation mode for the models. Defaults to "default".
  • model_version (float, optional: Version of the model to use. Defaults to the latest available version. Can also be set to 4.0, 3.0, 2.0, or 1.0 for older models.

Legacy Models (v1 to v3)

If you need to use legacy model versions (v1.0, v2.0, or v3.0), you must install the fastai dependency. These older models were built using fastai + timm, while the current v4+ models use segmentation-models-pytorch + timm.

See the model changelog for details on the differences between model versions.

To install with legacy model support:

pip install omnicloudmask[legacy]
uv add omnicloudmask --extra legacy
conda install conda-forge::omnicloudmask conda-forge::fastai

Once installed, you can specify the model version in your prediction functions:

from omnicloudmask import predict_from_array

# Use a legacy model version
pred_mask = predict_from_array(input_array, model_version=3.0)

Contributing

Contributions are welcome! Please submit a pull request or open an issue to discuss any changes.

License

This project is licensed under the MIT License

Acknowledgements

  • Special thanks to the CloudSen12 project for the dataset used for model versions 1.0, 2.0, 3.0 and 4.0.
  • Special thanks to the KappaSet authors for the dataset used for model versions 3.0 and 4.0.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

omnicloudmask-1.7.0.tar.gz (34.2 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

omnicloudmask-1.7.0-py3-none-any.whl (24.2 kB view details)

Uploaded Python 3

File details

Details for the file omnicloudmask-1.7.0.tar.gz.

File metadata

  • Download URL: omnicloudmask-1.7.0.tar.gz
  • Upload date:
  • Size: 34.2 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: uv/0.8.15

File hashes

Hashes for omnicloudmask-1.7.0.tar.gz
Algorithm Hash digest
SHA256 5583b9bb1e46469835bb61074c47a6037639e11323c9568bd87f059db4a7c1e9
MD5 8d3150fd94a17ca1b5e5cd20d47a7d62
BLAKE2b-256 2f20c33b33923abe1bb5965e699a6ffc9a5afc4a4efe72bad8cb614af9516698

See more details on using hashes here.

File details

Details for the file omnicloudmask-1.7.0-py3-none-any.whl.

File metadata

File hashes

Hashes for omnicloudmask-1.7.0-py3-none-any.whl
Algorithm Hash digest
SHA256 f18338b2af063e048f076de705fa64bca89231196eafdfbbf47cc0674b5ecfb4
MD5 bd46079c365d07098bbc5a7385b4f875
BLAKE2b-256 0222229bcaa7717a39235e2de4b013359894a569fb4c57c197dac412dc920bbb

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page