Skip to main content

CLI utils for working with a datacube index

Project description

odc.apps.dc_tools

Command line utilities for working with datacube index

Installation

pip install odc-apps-dc-tools

Usage

dc-sync-products

The tool dc-sync-products helps you by keeping a Datacube instance's list of products up to date with a CSV list of product names and definitions.

Basic usage is:

dc-sync-products <path-to-csv> --update-if-exists

The --update-if-exists flag is optional, and will update a product, including unsafe changes, if it already exists. The format for the CSV is as follows (note that you can have multiple products defined in one file):

product,definition
dem_srtm,https://raw.githubusercontent.com/digitalearthafrica/config/master/products/dem_srtm.odc-product.yaml
ls5_c2l2_sr;ls7_c2l2_sr;ls8_c2l2_sr;ls9_c2l2_sr,https://raw.githubusercontent.com/opendatacube/datacube-dataset-config/main/products/lsX_c2l2_sr.odc-product.yaml

dc-index-export-md

Metadata transformer

Simple usage:

TODO:

Extended usage:

TODO:

dc-index-from-tar

Index ODC metadata that is contained in a .tar file

Simple usage:

dc-index-from-tar 'path/to/file.tar'

Extended usage:

TODO:

sqs-to-dc

A tool to index from an SQS queue

Simple usage:

sqs-to-dc example-queue-name 'product-name-a product-name-b'

Extended usage:

Usage: sqs-to-dc [OPTIONS] QUEUE_NAME PRODUCT

  Iterate through messages on an SQS queue and add them to datacube

Options:
  --skip-lineage                  Default is not to skip lineage. Set to skip
                                  lineage altogether.

  --fail-on-missing-lineage / --auto-add-lineage
                                  Default is to fail if lineage documents not
                                  present in the database. Set auto add to try
                                  to index lineage documents.

  --verify-lineage                Default is no verification. Set to verify
                                  parent dataset definitions.

  --stac                          Expect STAC 1.0 metadata and attempt to
                                  transform to ODC EO3 metadata

  --odc-metadata-link TEXT        Expect metadata doc with ODC EO3 metadata
                                  link. Either provide '/' separated path to
                                  find metadata link in a provided metadata
                                  doc e.g. 'foo/bar/link', or if metadata doc
                                  is STAC, provide 'rel' value of the 'links'
                                  object having metadata link. e.g. 'STAC-
                                  LINKS-REL:odc_yaml'

  --limit INTEGER                 Stop indexing after n datasets have been
                                  indexed.

  --update                        If set, update instead of add datasets
  --update-if-exists              If the dataset already exists, update it
                                  instead of skipping it.

  --archive                       If set, archive datasets
  --allow-unsafe                  Allow unsafe changes to a dataset. Take
                                  care!

  --record-path TEXT              Filtering option for s3 path, i.e.
                                  'L2/sentinel-2-nrt/S2MSIARD/*/*/ARD-
                                  METADATA.yaml'

  --region-code-list-uri TEXT     A path to a list (one item per line, in txt
                                  or gzip format) of valide region_codes to
                                  include

  --absolute                      Use absolute paths when converting from stac

  --archive-less-mature           Find less mature versions of the dataset and
                                  archive them
                                  
  --publish-action SNS ARN        Publish indexing action to SNS topic

  --help                          Show this message and exit.

s3-to-dc

A tool for indexing from S3.

Simple usage:

s3-to-dc 's3://bucket/path/**/*.yaml' 'product-name-a product-name-b'

Extended usage:

The following command updates the datasets instead of adding them and allows unsafe changes. Be careful!

Usage: s3-to-dc [OPTIONS] URI PRODUCT

  Iterate through files in an S3 bucket and add them to datacube

Options:
  --skip-lineage                  Default is not to skip lineage. Set to skip
                                  lineage altogether.

  --fail-on-missing-lineage / --auto-add-lineage
                                  Default is to fail if lineage documents not
                                  present in the database. Set auto add to try
                                  to index lineage documents.

  --verify-lineage                Default is no verification. Set to verify
                                  parent dataset definitions.

  --stac                          Expect STAC 1.0 metadata and attempt to
                                  transform to ODC EO3 metadata

  --update                        If set, update instead of add datasets
  --update-if-exists              If the dataset already exists, update it
                                  instead of skipping it.

  --allow-unsafe                  Allow unsafe changes to a dataset. Take
                                  care!

  --skip-check                    Assume file exists when listing exact file
                                  rather than wildcard.

  --no-sign-request               Do not sign AWS S3 requests
  --request-payer                 Needed when accessing requester pays public
                                  buckets

  --archive-less-mature           Find less mature versions of the dataset and
                                  archive them

  --publish-action SNS ARN        Publish indexing action to SNS topic

  --help                          Show this message and exit.

thredds-to-dc

Index from a THREDDS server

Simple usage:

TODO:

Extended usage:

TODO:

esri-lc-to-dc

Removed, use the stac-to-dc tool instead.

  stac-to-dc \
    --catalog-href=https://planetarycomputer.microsoft.com/api/stac/v1/ \
    --collections='io-lulc'

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

odc_apps_dc_tools-1.9.8.tar.gz (51.7 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

odc_apps_dc_tools-1.9.8-py3-none-any.whl (55.0 kB view details)

Uploaded Python 3

File details

Details for the file odc_apps_dc_tools-1.9.8.tar.gz.

File metadata

  • Download URL: odc_apps_dc_tools-1.9.8.tar.gz
  • Upload date:
  • Size: 51.7 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.12

File hashes

Hashes for odc_apps_dc_tools-1.9.8.tar.gz
Algorithm Hash digest
SHA256 978ae5d80894c9469921e2b17beaf95440f2c45438fb57940b185fa858a880ce
MD5 f9071a967a3b2d2bcbfeddeceba9e59a
BLAKE2b-256 ec8f6ab12bcd57c9ab4673e51d99e832270d6d4734159538a5f93816ded5e7b3

See more details on using hashes here.

File details

Details for the file odc_apps_dc_tools-1.9.8-py3-none-any.whl.

File metadata

File hashes

Hashes for odc_apps_dc_tools-1.9.8-py3-none-any.whl
Algorithm Hash digest
SHA256 e082d2ece46d5a436a404e707b24ae86079261e175385fd27ed3f4914ef7428b
MD5 2739822980df4c0d0920d40db341b91d
BLAKE2b-256 b16281b2d94b52bd9fd1f7ff1b6a04a39770aee8fbc1b3e1f5d998e7252d97f2

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page