Skip to main content

A python package for building d4k microservcies

Project description

Introduction

This read me identifies a set of microservices that implement a range of functionality that allows for the handling of clinical trial data in an automated manner. The use cases addressed are

  1. Import of data held in a variety of data formats into study designs
  2. Export of data in a number of formats

The microservices within the system are:

  1. Registration Authority
  2. Clinical Recording Model
  3. Controlled Terminology
  4. Biomedical Concepts
  5. SDTM
  6. Forms
  7. Study

The services are all python services running on FastAPI and uvicorn. They can be run locally using a single shared Neo4j instance or deployed. When deployed each micorservices requires a separate Aura Neo4j instance.

Running Locally

Use a single Neo4 instance will all databases for the microservices setup within one instance. Typical names for a devlepoment environment are

  • ra-service-dev
  • crm-service-dev
  • bc-service-dev
  • sdtm-service-dev
  • study-service-dev
  • form-service-dev

Note: there are no separate data bases for Study Import service or the Study Data Import service

Port Numbers

User Interface Services

Microservice Port
CT 8000
BC 8001
Study 8002
Form 8003
SDTM 8004

Data Services

Microservice Port
RA 8010
CRM 8011
CT 8012
BC 8013
Study 8014
Study Import 8015
Form 8016
SDTM 8017
Study Data Import 8018

To Run

Use the microservice.command file to run. Note this file is mac specific. This will start all of the services and the UIs. There is also a dev_server.sh file in each git repo to run each element individually.

Deployed

Overview

To deploy each microservice the following actions should be followed.

  1. Create Aura Neo4j instance.
  2. Load the data using the data prep utility
  3. Deploy the microservice
  4. Check all running

Neo4j Aura

Use the Aura console to create the database as per the site instructions. Download the file holding the credentials (username and password) for the instance

Data Load

Overview

From the appropriate data prep project

  1. Setup the virtual environment
  2. Set the environment variables for the credentials of the neo4j instance
  3. Set production
  4. Run the load python file

Virtual Environment

Run . ./setup_env.sh

Environment Variables

Modify the .production_env file. The file will need to contain the following lines

NEO4J_URI=<database uri>
NEO4J_DB_NAME=neo4j
NEO4J_USERNAME=<username>
NEO4J_PASSWORD=<password>
GITHUB=<URL of the github repo main page>

The database URI will look something like neo4j+s://a1bc23d4.databases.neo4j.io The githb repo URL will look something like https://raw.githubusercontent.com/data4knowledge/ra_prep/main/

Set Production

We need to tell the load utility to use the production database rather than the local (development) one. This is done via setting the PYTHON_ENVIRONMENTenvironment variable. To so this run . ./set_production.sh

Load Data

There will be a python rpogram to load the data. It will be named stage_<n>_load.py. Run this program by entering python stage_2_load.py, here n is 2.

The output from the program should look somethig like

Deleting database ...
Database deleted. Load new data ...
https://raw.githubusercontent.com/data4knowledge/ra_prep/main/load_data/node-namespace-1.csv
https://raw.githubusercontent.com/data4knowledge/ra_prep/main/load_data/relationship-manages-1.csv
https://raw.githubusercontent.com/data4knowledge/ra_prep/main/load_data/node-registration_authority-1.csv
<Record file='progress.csv' source='file' format='csv' nodes=9 relationships=7 properties=58 time=4294 rows=0 batchSize=-1 batches=0 done=True data=None>
Load complete. 9 nodes and 7 relationships loaded in 4294 milliseconds.

Deploy Micorservice

The microservices are deployed using the fly.io cloud service.

  1. General python instructions are available here https://fly.io/docs/languages-and-frameworks/python/
  2. Installing the fly command line tool is detailed here https://fly.io/docs/hands-on/install-flyctl/
  3. deploy the app
  4. Set environment variables

Environment Variables

Setting environment variables on the server is achieved by using th ecommand line program, either one at a time

fly secrets set SUPER_SECRET_KEY=password1234

or multiple values, note space delimited

fly secrets set NEO4J_URI=xxx NEO4J_PASSWORD=yyy

Loading Studies & Data

To be defined.

Building ppackage

Use pip to install build and twine. use the following commands to build python -m build and upload twine upload dist/* to pypi.org

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

d4kms-0.2.tar.gz (19.9 kB view details)

Uploaded Source

Built Distribution

d4kms-0.2-py3-none-any.whl (21.4 kB view details)

Uploaded Python 3

File details

Details for the file d4kms-0.2.tar.gz.

File metadata

  • Download URL: d4kms-0.2.tar.gz
  • Upload date:
  • Size: 19.9 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.0.0 CPython/3.10.1

File hashes

Hashes for d4kms-0.2.tar.gz
Algorithm Hash digest
SHA256 a22db5cfd0584689e76c2ba47ef9e4c30126a6ba97e94794174d1928c7a60de1
MD5 93ec65fe4722bab9309ebceef6cb030a
BLAKE2b-256 eecc7878ff4021442c086b2cf7d475088af9df5896a9f33a08ebd3edace49131

See more details on using hashes here.

File details

Details for the file d4kms-0.2-py3-none-any.whl.

File metadata

  • Download URL: d4kms-0.2-py3-none-any.whl
  • Upload date:
  • Size: 21.4 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.0.0 CPython/3.10.1

File hashes

Hashes for d4kms-0.2-py3-none-any.whl
Algorithm Hash digest
SHA256 41c138f98229fa2b2169e9f3544f3ec6f8a13c70ba1d53c090884763460f8dba
MD5 a74a3715234d90f781818a6b6fcb31cd
BLAKE2b-256 4f13176de663773eb36512158059aa9096a3c26b132ca387f2547b225b860cea

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page