Skip to main content

A map binning tool for spatial resampling

Project description

Map Binning Tool

Python Version License: MIT CI/CD Pipeline codecov

A Python package for spatial resampling and binning of geospatial data, specifically designed for oceanographic datasets. This tool enables efficient downsampling of high-resolution gridded data onto coarser grids while preserving spatial accuracy through intelligent neighborhood averaging.

Overview

The Map Binning Tool provides a robust solution for spatial data aggregation, particularly useful for:

  • Downsampling high-resolution oceanographic data (e.g., sea level anomaly, ocean currents)
  • Creating consistent multi-resolution datasets
  • Reducing computational load while maintaining spatial representativeness
  • Processing time-series of gridded data efficiently

The package uses k-d tree algorithms for fast spatial queries and supports both in-memory processing and persistent caching of spatial indices for repeated operations.

Key Features

  • Efficient Spatial Binning: Uses scipy's cKDTree for fast nearest-neighbor searches
  • Flexible Grid Support: Works with any xarray-compatible gridded dataset
  • Automatic Radius Calculation: Intelligently determines search radius based on target grid spacing
  • Persistent Caching: Save and reuse spatial indices using pickle serialization
  • Time Series Support: Handles datasets with temporal dimensions
  • Memory Efficient: Processes large datasets without excessive memory usage
  • Oceanographic Focus: Optimized for CMEMS and similar oceanographic data formats

Installation

From PyPI (Recommended)

pip install map-binning

With optional dependencies for development

pip install map-binning[dev]

Developer Installation

Using conda environment

# Create and activate conda environment
conda env create -f environment.yml
conda activate map-binning

# Install the package in development mode
pip install -e .

From source

git clone <repository-url>
cd map_binning
pip install -e .

Quick Start

Basic Usage

import xarray as xr
from map_binning import Binning

# Load your datasets
ds_high = xr.open_dataset('high_resolution_data.nc')
ds_low = xr.open_dataset('low_resolution_grid.nc')

# Initialize the binning tool
binning = Binning(
    ds_high=ds_high,
    ds_low=ds_low,
    var_name='sla',  # variable in the dataset to bin (e.g., sea level anomaly)
    xdim_name='longitude',  # longitude dimension name
    ydim_name='latitude',   # latitude dimension name
    search_radius=0.1  # optional: search radius in degrees
)

# Perform binning
result = binning.mean_binning()

Advanced Usage with Caching

# Create binning index and save it for reuse
result = binning.mean_binning(
    precomputed_binning_index=False,
    pickle_filename="my_binning_index.pkl",
    pickle_location="./cache"
)

# Reuse the saved index for subsequent operations
result = binning.mean_binning(
    precomputed_binning_index=True,
    pickle_filename="my_binning_index.pkl",
    pickle_location="./cache"
)

API Reference

Binning Class

Constructor Parameters

  • ds_high (xr.Dataset): High-resolution source dataset
  • ds_low (xr.Dataset): Low-resolution target grid dataset
  • var_name (str): Name of the variable to bin
  • xdim_name (str, optional): Longitude dimension name (default: 'lon')
  • ydim_name (str, optional): Latitude dimension name (default: 'lat')
  • search_radius (float, optional): Search radius in degrees (auto-calculated if None)

Methods

create_binning_index() Creates a spatial mapping between high and low resolution grids.

mean_binning(precomputed_binning_index=False, pickle_filename=None, pickle_location=None) Performs spatial binning using mean aggregation.

Parameters:

  • precomputed_binning_index (bool): Use pre-saved spatial index
  • pickle_filename (str): Filename for saving/loading spatial index
  • pickle_location (str): Directory path for pickle files

Returns: xr.DataArray with binned data on the target grid

Project Structure

map_binning/
├── map_binning/           # Main package directory
│   ├── __init__.py        # Package initialization
│   ├── binning.py         # Core binning algorithms
│   ├── index_store.py     # Pickle serialization utilities
│   └── main.py            # Command-line interface
├── notebooks/             # Jupyter notebooks for examples
│   └── cmems_nrt_coastal_bin.ipynb
├── tests/                 # Unit tests
│   ├── __init__.py
│   └── test_main.py
├── pickle_folder/         # Default location for cached indices
├── pyproject.toml         # Project configuration
├── environment.yml        # Conda environment specification
├── .env.template          # Environment variables template
└── README.md              # This file

Configuration for CMEMS data download

Environment Variables

Copy .env.template to .env and configure:

# Copernicus Marine Service credentials (if using CMEMS data)
COPERNICUSMARINE_SERVICE_USERNAME=<your_username>
COPERNICUSMARINE_SERVICE_PASSWORD=<your_password>

Dependencies

Core dependencies:

  • numpy: Numerical computing
  • scipy: Scientific computing (k-d tree algorithms)
  • xarray: Labeled multi-dimensional arrays
  • netcdf4: NetCDF file I/O
  • python-dotenv: Environment variable management

Development dependencies:

  • pytest: Unit testing framework
  • black: Code formatting
  • flake8: Code linting
  • mypy: Static type checking

Examples

Working with CMEMS Data

import xarray as xr
from map_binning import Binning

# Example with Copernicus Marine data
ds_high = xr.open_dataset('cmems_high_res_sla.nc')
ds_low = xr.open_dataset('cmems_low_res_grid.nc')

# Initialize for sea level anomaly processing
sla_binning = Binning(
    ds_high=ds_high,
    ds_low=ds_low,
    var_name='sla',
    xdim_name='longitude',
    ydim_name='latitude'
)

# Process and cache the result
binned_sla = sla_binning.mean_binning(
    pickle_filename="cmems_sla_index.pkl",
    pickle_location="./cache"
)

# Save the result
binned_sla.to_netcdf('binned_sla_data.nc')

Time Series Processing

The tool automatically handles time dimensions:

# Works seamlessly with time-varying datasets
# Input: (time, lat, lon) -> Output: (time, lat_low, lon_low)
result = binning.mean_binning()

Performance Considerations

  • Memory Usage: The tool processes data in chunks and uses efficient numpy operations
  • Spatial Index Caching: Save computed spatial indices to avoid recalculation
  • Grid Resolution: Performance scales with the product of grid sizes
  • Search Radius: Smaller radii improve performance but may miss relevant data points

Contributing

We welcome contributions! Please follow these steps:

  1. Fork the repository
  2. Create a feature branch (git checkout -b feature/amazing-feature)
  3. Make your changes and add tests
  4. Run the test suite (pytest)
  5. Format your code (black map_binning/)
  6. Submit a pull request

Development Setup

# Clone and setup development environment
git clone <repository-url>
cd map-binning-project
conda env create -f environment.yml
conda activate map-binning
pip install -e .[dev]

# Run tests
pytest

# Format code
black map_binning/

# Type checking
mypy map_binning/

Testing

# Run all tests
pytest

# Run with coverage
pytest --cov=map_binning

# Run specific test file
pytest tests/test_main.py

License

This project is licensed under the MIT License - see the LICENSE file for details.

Citation

If you use this tool in your research, please cite:

@software{map_binning_2024,
  author = {Chia-Wei Hsu},
  title = {Map Binning Tool: Spatial Resampling for Oceanographic Data},
  url = {https://github.com/chiaweh2/map_binning},
  version = {0.1.0},
  year = {2024}
}

Support

  • Issues: Please report bugs and feature requests via GitHub Issues
  • Documentation: Additional examples available in the notebooks/ directory
  • Contact: Chia-Wei Hsu (chiaweh2@uci.edu)

Acknowledgments

  • Built with support for Copernicus Marine Environment Monitoring Service (CMEMS) data
  • Utilizes scipy's efficient spatial algorithms
  • Designed for the oceanographic research community

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

map_binning-0.2.1.tar.gz (17.4 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

map_binning-0.2.1-py3-none-any.whl (9.0 kB view details)

Uploaded Python 3

File details

Details for the file map_binning-0.2.1.tar.gz.

File metadata

  • Download URL: map_binning-0.2.1.tar.gz
  • Upload date:
  • Size: 17.4 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.11.13

File hashes

Hashes for map_binning-0.2.1.tar.gz
Algorithm Hash digest
SHA256 928a201f23b0c85ba481d026edfd22b96520305e8f1834e2c18022c4412bc806
MD5 3746d9949db925f474e20b7c85e39b25
BLAKE2b-256 ece75d53aaa11eb9f84f8cca066cc903608e2a1c981a6d9331bbcb070238937b

See more details on using hashes here.

File details

Details for the file map_binning-0.2.1-py3-none-any.whl.

File metadata

  • Download URL: map_binning-0.2.1-py3-none-any.whl
  • Upload date:
  • Size: 9.0 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.11.13

File hashes

Hashes for map_binning-0.2.1-py3-none-any.whl
Algorithm Hash digest
SHA256 b1979c6b34479593349e357cad8866d7f4fe4d4fafd5cf5e18627b76c33fe112
MD5 8c769bd87476aac5022dba3b6e80a4e5
BLAKE2b-256 73c42d1c3060db056f1c15dba4baf96c48fed406f34d624a033facefd971bf85

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page