Skip to main content

Package for handling bioinformatics data automations between data sources

Project description

bioforklift

Tests

Automation Data Movement and Integration Library for Sample Datastores

🏗️🏗️🏗️🏗️🏗️🏗️🏗️🏗️🏗️🏗️🏗️🏗️🏗️🏗️🏗️🏗️🏗️🏗️🏗️🏗️🏗️🏗️🏗️🏗️🏗️🏗️🏗️🏗️🏗️

🚧 Under Construction 🚧

Getting Setup

This project uses poetry for project management

If you don't have poetry present, please install it with: pip install poetry

Then run poetry env activate which will create your environment: poetry env activate

Next, install the dependencies listed in poetry.lock utilizing: poetry install

The dependencies will be installed based on the locked versions in the poetry.lock file, since I already ran poetry install and pushed the lock file. For more information on poetry, read here: https://python-poetry.org/docs/basic-usage/

Finally, re authorize your gcloud authentication. This obtains your credentials via a web flow and stores them in 'the well-known location for Application Default Credentials'. Now any code/SDK you run will be able to find the credentials automatically. This is a good stand-in when you want to locally test code which would normally run on a server and use a server-side credentials file. gcloud auth application-default login

Note

This is a first time dump of everything I've been putting together for an automation library for our data movement needs

Overview

bioforklift Base Architecture

TODO:

  • Add target workspace entry for Terra class
  • Add test suite for bigquery layer
  • Add Terra2Bq integration layer
  • Add module level logging and better error handling
  • Define key yaml tags with team
  • Test scope of bigquery range
  • Test, Test, Test

Biggest lift to do is scope out what we actually want to include for the bigquery samples class and how we want to name key identifiers in the yamls, develops some internal schema for that, and then after that we should be flying.

🥶

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

bioforklift-0.1.0.tar.gz (62.0 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

bioforklift-0.1.0-py3-none-any.whl (72.3 kB view details)

Uploaded Python 3

File details

Details for the file bioforklift-0.1.0.tar.gz.

File metadata

  • Download URL: bioforklift-0.1.0.tar.gz
  • Upload date:
  • Size: 62.0 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/2.1.1 CPython/3.10.12 Linux/6.8.0-1026-gcp

File hashes

Hashes for bioforklift-0.1.0.tar.gz
Algorithm Hash digest
SHA256 3205a4cdf6d4f1f736e1eb9209d6badf8e4b63467d32ab4bd8d9f26902bff08f
MD5 36d2f684a8bdc649d03826594bae212b
BLAKE2b-256 78d58f5a2d83ea0441c9479ff6c5e1fd6b3c4fc4d53aaba1d0b4880e025acd3c

See more details on using hashes here.

File details

Details for the file bioforklift-0.1.0-py3-none-any.whl.

File metadata

  • Download URL: bioforklift-0.1.0-py3-none-any.whl
  • Upload date:
  • Size: 72.3 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/2.1.1 CPython/3.10.12 Linux/6.8.0-1026-gcp

File hashes

Hashes for bioforklift-0.1.0-py3-none-any.whl
Algorithm Hash digest
SHA256 dbd196b37bf1d074b87b44f1a7a80a3e2e8f621de165e236499a05598d23e13c
MD5 53d219ed2677bc727bfc62507b0e0eec
BLAKE2b-256 65e7cb3a9dd45f10c70b077e9c570817c5fa4869ff5bf11c2c27009ea8f4dbe0

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page