Skip to main content

Enumeration and ops library for the OPERA DIST-S1 project

Project description

dist-s1-enumerator

PyPI license PyPI pyversions PyPI version Conda version Conda platforms

This is a Python library for enumerating OPERA RTC-S1 inputs necessary for the creation of OPERA DIST-S1 products. The library can enumerate inputs for the creation of a single DIST-S1 product or a time-series of DIST-S1 products over a large area spanning multiple passes. The DIST-S1 measures disturbance comparing a baseline of RTC-S1 images (pre-images) to a current set of acquisition images (post-images). This library also provides functionality for downloading the OPERA RTC-S1 data from ASF DAAC.

Installation/Setup

We recommend managing dependencies and virutal environments using mamba/conda.

mamba update -f environment.yml  # creates a new environment dist-s1-enumerator
conda activate dist-s1-enumerator
pip install dist-s1-enumerator
python -m ipykernel install --user --name dist-s1-enumerator

Downloading data

For searching through the metadata of OPERA RTC-S1, you will not need any earthdata credentials. For downloading data from the ASF DAAC, you will need to make sure you have a Earthdata credentials (see: https://urs.earthdata.nasa.gov/) and successfully accepted the ASF terms of use (this can be checked by downloading any product at the ASF DAAC using your Earthdata credentials: https://search.asf.alaska.edu/). You will need to create or append to ~/.netrc file with these credentials:

machine urs.earthdata.nasa.gov
    login <your_username>
    password <your_password>

Development installation

Same as above replacing pip install dist-s1-enumerator with pip install -e ..

Usage

For triggering DIST-S1 Workflows

workflow_inputs = enumerate_dist_s1_workflow_inputs(mgrs_tile_ids='19HBD',
                                                    track_numbers=None,
                                                    start_acq_dt='2023-11-01',
                                                    stop_acq_dt='2024-04-01',
                                                    lookback_strategy='multi_window',
                                                    delta_lookback_days=365,
                                                    max_pre_imgs_per_burst=5)

Yields:

[{'mgrs_tile_id': '19HBD', 'post_acq_date': '2023-11-05', 'track_number': 91},
 {'mgrs_tile_id': '19HBD', 'post_acq_date': '2023-11-10', 'track_number': 156},
 {'mgrs_tile_id': '19HBD', 'post_acq_date': '2023-11-12', 'track_number': 18}...]

Where these fields uniquely determine a DIST-S1 product and can be used to trigger the workflow.

For collecting DIST-S1 inputs

The above example tells us the recent acquisition date that disturbance is made relative to (post_acq_date) over an MGRS tile (mgrs_tile_id). However, there are many OPERA RTC-S1 products used on that given date and to establish a baseline. To enumerate all the necessary inputs (which can be further localized with this library), see the Jupyter notebooks.

Identifiers for DIST-S1 products

Of course, knowing all the OPERA RTC-S1 products (pre-images and post-images) necessary for a DIST-S1 product uniquely identifies the products. However, all these inputs can be amount to upwards of 100 products for each DIST-S1 product and is not human parsable. Thus, it is helpful to know alterate ways to identify and trigger the DIST-S1 product and its' workflow.

Altenrately, we can uniqely identify a DIST-S1 product via the following fields:

  1. MGRS Tile ID
  2. Track Number
  3. Post-image acquisition time (within 1 day)

As shown in For triggering DIST-S1 Workflows section, that is precisely the output of enumerate_dist_s1_workflow_inputs.

We now explain why these fields uniquely identify DIST-S1 products. Each DIST-S1 product is resampled to an MGRS tile. One might assume that the post-image acquisition time is enough - however, there are particular instances when Sentinel-1 A and Sentinel-1 C will pass each other in the same day and so fixing the track number differentiates between the two sets of acquisired imagery. Thus, it is important to specify the date in addition to the track number. In theory, we could specify the exact time of acquisition, but we have elected to use track numbers. It is also important to note that we are assuming the selection of pre-images (once a post-image set is selected) is fixed. Indeed, varying a baseline of pre-images by which to measure disturbance will alter the final DIST-S1 product. Indeed, we can modify strategies of pre-image selection using this library (e.g. multi_window vs. immediate_lookback), but for DIST-S1 generation which has a fixed strategy with associated parameters, the above 3 fields uniquely identify a DIST-S1 product.

Testing

For the test suite:

  1. Install papermill via conda-forge (currently not supported by 3.13)
  2. Run pytest tests

There are two category of tests: unit tests and integration tests. The former can be run using pytest tests -m 'not integration' and similarly the latter with pytest tests -m 'integration'. The intgeration tests are those that can be integrated into the DAAC data access workflows and thus require internet access with earthdata credentials setup correctly (as described above). The unit tests mock the necessary data inputs. The integration tests that are the most time consuming are represented by the notebooks and are run only upon a release PR. These notebook tests are tagged with notebooks and can be excluded from the other tests with pytest tests -m 'not notebooks'.

Contributing

We welcome contributions to this open-source package. To do so:

  1. Create an GitHub issue ticket desrcribing what changes you need (e.g. issue-1)
  2. Fork this repo
  3. Make your modifications in your own fork
  4. Make a pull-request (PR) in this repo with the code in your fork and tag the repo owner or a relevant contributor.

We use ruff and associated linting packages to ensure some basic code quality (see the environment.yml). These will be checked for each commit in a PR. Try to write tests wherever possible.

Support

  1. Create an GitHub issue ticket desrcribing what changes you would like to see or to report a bug.
  2. We will work on solving this issue (hopefully with you).

Acknowledgements

See the LICENSE file for copyright information.

This package was developed as part of the Observational Products for End-Users from Remote Sensing Analysis (OPERA) project. This work was originally carried out at the Jet Propulsion Laboratory, California Institute of Technology, under a contract with the National Aeronautics and Space Administration (80NM0018D0004). Copyright 2024 by the California Institute of Technology. United States Government Sponsorship acknowledged.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

dist_s1_enumerator-1.0.2.tar.gz (30.8 MB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

dist_s1_enumerator-1.0.2-py3-none-any.whl (28.4 MB view details)

Uploaded Python 3

File details

Details for the file dist_s1_enumerator-1.0.2.tar.gz.

File metadata

  • Download URL: dist_s1_enumerator-1.0.2.tar.gz
  • Upload date:
  • Size: 30.8 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for dist_s1_enumerator-1.0.2.tar.gz
Algorithm Hash digest
SHA256 a4d66f4822e106f96be43209a7f723f6a3f0d398c5d3edf87b572671637e2bba
MD5 168da116a938f92e26ceb124c5fff4a8
BLAKE2b-256 7eef3226efdee145f9312f26b6f04bc7d4f415142e7e83d976ef1eefb31cbd60

See more details on using hashes here.

File details

Details for the file dist_s1_enumerator-1.0.2-py3-none-any.whl.

File metadata

File hashes

Hashes for dist_s1_enumerator-1.0.2-py3-none-any.whl
Algorithm Hash digest
SHA256 e3e9061858ac6b37c0a5a8843517694a04df8f81265a71b6c35457c7914208c3
MD5 99ca9f3bab68859561da0a963fa1f4dc
BLAKE2b-256 fee1b4e3c562cc7da693cf1d387e5947ce3450305807fa275adb5e00685955d5

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page