Skip to main content

A library that defines AIND data schema and validates JSON files.

Project description

aind-data-schema

License Code Style

A library that defines AIND data schema and validates JSON files.

Overview

This repository contains the schemas needed to ingest and validate metadata that are essential to ensuring AIND data collection is completely reproducible. Our general approach is to semantically version core schema classes and include those version numbers in serialized metadata so that we can flexibly evolve the schemas over time without requiring difficult data migrations. In the future, we will provide a browsable list of these classes rendered to JSONschema, including all historic versions.

Be aware that this package is still under heavy preliminary development. Expect breaking changes regularly, although we will communicate these through semantic versioning.

A simple example:

from aind_data_schema import Subject
import datetime

t = datetime.datetime(2022, 11, 22, 8, 43, 00)

s = Subject(
    species="Mus musculus",
    subject_id="12345",
    sex="Male",
    date_of_birth=t.date(),
    genotype="Emx1-IRES-Cre;Camk2a-tTA;Ai93(TITL-GCaMP6f)",
    home_cage_enrichment="other",
    background_strain="C57BL/6J",
)

with open("subject.json", "w") as f:
    f.write(s.json(indent=3))
{
   "describedBy": "https://raw.githubusercontent.com/AllenNeuralDynamics/aind-data-schema/main/src/aind_data_schema/subject.py",
   "schema_version": "0.2.2",
   "species": "Mus musculus",
   "subject_id": "12345",
   "sex": "Male",
   "date_of_birth": "2022-11-22",
   "genotype": "Emx1-IRES-Cre;Camk2a-tTA;Ai93(TITL-GCaMP6f)",
   "mgi_allele_ids": null,
   "background_strain": "C57BL/6J",
   "source": null,
   "rrid": null,
   "restrictions": null,
   "breeding_group": null,
   "maternal_id": null,
   "maternal_genotype": null,
   "paternal_id": null,
   "paternal_genotype": null,
   "light_cycle": null,
   "home_cage_enrichment": "other",
   "wellness_reports": null,
   "notes": null
}

Installing and Upgrading

To install the latest version:

pip install aind-data-schema

Every merge to the main branch is automatically tagged with a new major/minor/patch version and uploaded to PyPI. To upgrade to the latest version:

pip install aind-data-schema --upgrade

To develop the code, check out this repo and run the following in the cloned directory:

pip install -e .[dev]

Contributing

If you've found a bug in the schemas or would like to make a minor change, open an Issue on this repository. If you'd like to propose a large change or addition, or generally have a question about how things work, head start a new Discussion!

Linters and testing

There are several libraries used to run linters, check documentation, and run tests.

  • Please test your changes using the coverage library, which will run the tests and log a coverage report:
coverage run -m unittest discover && coverage report
  • Use interrogate to check that modules, methods, etc. have been documented thoroughly:
interrogate .
  • Use flake8 to check that code is up to standards (no unused imports, etc.):
flake8 .
  • Use black to automatically format the code into PEP standards:
black .
  • Use isort to automatically sort import statements:
isort .

Pull requests

For internal members, please create a branch. For external members, please fork the repo and open a pull request from the fork. We'll primarily use Angular style for commit messages. Roughly, they should follow the pattern:

<type>(<scope>): <short summary>

where scope (optional) describes the packages affected by the code changes and type (mandatory) is one of:

  • build: Changes that affect the build system or external dependencies (example scopes: pyproject.toml, setup.py)
  • ci: Changes to our CI configuration files and scripts (examples: .github/workflows/ci.yml)
  • docs: Documentation only changes
  • feat: A new feature
  • fix: A bug fix
  • perf: A code change that improves performance
  • refactor: A code change that neither fixes a bug nor adds a feature
  • test: Adding missing tests or correcting existing tests

Documentation

To generate the rst files source files for documentation, run:

sphinx-apidoc -o doc_template/source/ src 

Then to create the documentation html files, run:

sphinx-build -b html doc_template/source/ doc_template/build/html

More info on sphinx installation can be found here: https://www.sphinx-doc.org/en/master/usage/installation.html

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

aind-data-schema-0.9.2.tar.gz (46.2 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

aind_data_schema-0.9.2-py3-none-any.whl (29.8 kB view details)

Uploaded Python 3

File details

Details for the file aind-data-schema-0.9.2.tar.gz.

File metadata

  • Download URL: aind-data-schema-0.9.2.tar.gz
  • Upload date:
  • Size: 46.2 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.1 CPython/3.9.16

File hashes

Hashes for aind-data-schema-0.9.2.tar.gz
Algorithm Hash digest
SHA256 ecc4c64aacd11cf17bb9f6729eb65b91f329f1f8fa0cedd8534f1e5bb3cd63c3
MD5 5c70b0a01cc8c1d37d543e1184c1b796
BLAKE2b-256 dce466e2602c4a21721461d7fa035b0e4d134be4c3ddc74f1cabb5f633dcd71c

See more details on using hashes here.

File details

Details for the file aind_data_schema-0.9.2-py3-none-any.whl.

File metadata

File hashes

Hashes for aind_data_schema-0.9.2-py3-none-any.whl
Algorithm Hash digest
SHA256 158277a3ac93b8c18a7a4f4952a1a71fe6c58d16666bcba4a074f7d099df4a58
MD5 53a16dd445b3709cb724d59f666ae494
BLAKE2b-256 baf22a8d488185d466814f66586640ab36351d49eddb26d741eacfa4c2628948

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page