OSS Implementation of a DataJunction Query Service
Project description
DataJunction Query Service
This repository (DJQS) is an open source implementation of a DataJunction query service. It allows you to create catalogs and engines that represent sqlalchemy connections. Configuring a DJ server to use a DJQS server allows DJ to query any of the database technologies supported by sqlalchemy.
Quickstart
To get started, clone this repo and start up the docker compose environment.
git clone https://github.com/DataJunction/djqs
cd djqs
docker compose up
Creating Catalogs
Catalogs can be created using the POST /catalogs/
endpoint.
curl -X 'POST' \
'http://localhost:8001/catalogs/' \
-H 'accept: application/json' \
-H 'Content-Type: application/json' \
-d '{
"name": "djdb"
}'
Creating Engines
Engines can be created using the POST /engines/
endpoint.
curl -X 'POST' \
'http://localhost:8001/engines/' \
-H 'accept: application/json' \
-H 'Content-Type: application/json' \
-d '{
"name": "sqlalchemy-postgresql",
"version": "15.2",
"uri": "postgresql://dj:dj@postgres-roads:5432/djdb"
}'
Engines can be attached to existing catalogs using the POST /catalogs/{name}/engines/
endpoint.
curl -X 'POST' \
'http://localhost:8001/catalogs/djdb/engines/' \
-H 'accept: application/json' \
-H 'Content-Type: application/json' \
-d '[
{
"name": "sqlalchemy-postgresql",
"version": "15.2"
}
]'
Executing Queries
Queries can be submitted to DJQS for a specified catalog and engine.
curl -X 'POST' \
'http://localhost:8001/queries/' \
-H 'accept: application/json' \
-H 'Content-Type: application/json' \
-d '{
"catalog_name": "djdb",
"engine_name": "sqlalchemy-postgresql",
"engine_version": "15.2",
"submitted_query": "SELECT * from roads.repair_orders",
"async_": false
}'
Async queries can be submitted as well.
curl -X 'POST' \
'http://localhost:8001/queries/' \
-H 'accept: application/json' \
-H 'Content-Type: application/json' \
-d '{
"catalog_name": "djdb",
"engine_name": "sqlalchemy-postgresql",
"engine_version": "15.2",
"submitted_query": "SELECT * from roads.repair_orders",
"async_": true
}'
response
{
"catalog_name": "djdb",
"engine_name": "sqlalchemy-postgresql",
"engine_version": "15.2",
"id": "<QUERY ID HERE>",
"submitted_query": "SELECT * from roads.repair_orders",
"executed_query": null,
"scheduled": null,
"started": null,
"finished": null,
"state": "ACCEPTED",
"progress": 0,
"results": [],
"next": null,
"previous": null,
"errors": []
}
The query id provided in the response can then be used to check the status of the running query and get the results once it’s completed.
curl -X 'GET' \
'http://localhost:8001/queries/<QUERY ID HERE>/' \
-H 'accept: application/json'
response
{
"catalog_name": "djdb",
"engine_name": "sqlalchemy-postgresql",
"engine_version": "15.2",
"id": "$QUERY_ID",
"submitted_query": "SELECT * from roads.repair_orders",
"executed_query": "SELECT * from roads.repair_orders",
"scheduled": "2023-02-28T07:27:55.367162",
"started": "2023-02-28T07:27:55.367387",
"finished": "2023-02-28T07:27:55.502412",
"state": "FINISHED",
"progress": 1,
"results": [
{
"sql": "SELECT * from roads.repair_orders",
"columns": [...],
"rows": [...],
"row_count": 25
}
],
"next": null,
"previous": null,
"errors": []
}
Reflection
If running a [reflection service](https://github.com/DataJunction/djrs), that service can leverage the
POST /table/{table}/columns/
endpoint of DJQS to get column names and types for a given table.
curl -X 'GET' \
'http://localhost:8001/table/djdb.roads.repair_orders/columns/?engine=sqlalchemy-postgresql&engine_version=15.2' \
-H 'accept: application/json'
response
{
"name": "djdb.roads.repair_orders",
"columns": [
{
"name": "repair_order_id",
"type": "INT"
},
{
"name": "municipality_id",
"type": "STR"
},
{
"name": "hard_hat_id",
"type": "INT"
},
{
"name": "order_date",
"type": "DATE"
},
{
"name": "required_date",
"type": "DATE"
},
{
"name": "dispatched_date",
"type": "DATE"
},
{
"name": "dispatcher_id",
"type": "INT"
}
]
}
DuckDB
DJQS includes an example of using DuckDB as an engine and it comes preloaded with the roads example database.
Create a djduckdb
catalog and a duckdb
engine.
curl -X 'POST' \
'http://localhost:8001/catalogs/' \
-H 'accept: application/json' \
-H 'Content-Type: application/json' \
-d '{
"name": "djduckdb"
}'
curl -X 'POST' \
'http://localhost:8001/engines/' \
-H 'accept: application/json' \
-H 'Content-Type: application/json' \
-d '{
"name": "duckdb",
"version": "0.7.1",
"uri": "duckdb://local[*]"
}'
curl -X 'POST' \
'http://localhost:8001/catalogs/djduckdb/engines/' \
-H 'accept: application/json' \
-H 'Content-Type: application/json' \
-d '[
{
"name": "duckdb",
"version": "0.7.1"
}
]'
Now you can submit DuckDB SQL queries.
curl -X 'POST' \
'http://localhost:8001/queries/' \
-H 'accept: application/json' \
-H 'Content-Type: application/json' \
-d '{
"catalog_name": "djduckdb",
"engine_name": "duckdb",
"engine_version": "0.7.1",
"submitted_query": "SELECT * FROM roads.us_states LIMIT 10",
"async_": false
}'
Spark
DJQS includes an example of using Spark as an engine. To try it, start up the docker compose environment and then load the example roads database into Spark.
docker exec -it djqs /bin/bash -c "python /code/docker/spark_load_roads.py"
Next, create a djspark
catalog and a spark
engine.
curl -X 'POST' \
'http://localhost:8001/catalogs/' \
-H 'accept: application/json' \
-H 'Content-Type: application/json' \
-d '{
"name": "djspark"
}'
curl -X 'POST' \
'http://localhost:8001/engines/' \
-H 'accept: application/json' \
-H 'Content-Type: application/json' \
-d '{
"name": "spark",
"version": "3.3.2",
"uri": "spark://local[*]"
}'
curl -X 'POST' \
'http://localhost:8001/catalogs/djspark/engines/' \
-H 'accept: application/json' \
-H 'Content-Type: application/json' \
-d '[
{
"name": "spark",
"version": "3.3.2"
}
]'
Now you can submit Spark SQL queries.
curl -X 'POST' \
'http://localhost:8001/queries/' \
-H 'accept: application/json' \
-H 'Content-Type: application/json' \
-d '{
"catalog_name": "djspark",
"engine_name": "spark",
"engine_version": "3.3.2",
"submitted_query": "SELECT * FROM roads.us_states LIMIT 10",
"async_": false
}'
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file datajunction_query-0.0.1a64.tar.gz
.
File metadata
- Download URL: datajunction_query-0.0.1a64.tar.gz
- Upload date:
- Size: 162.4 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: python-httpx/0.27.2
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 42e0500a5d8adea32908a556798fa1e77e1057d5cde96fe66f805bf4536d9467 |
|
MD5 | 294c4ec005da4b30c4858e2e5bc60f77 |
|
BLAKE2b-256 | da1a2641fae14cc7a7e8439c13e7ab882be05b3a60121284eaf347a91bfaf0ef |
File details
Details for the file datajunction_query-0.0.1a64-py3-none-any.whl
.
File metadata
- Download URL: datajunction_query-0.0.1a64-py3-none-any.whl
- Upload date:
- Size: 24.3 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: python-httpx/0.27.2
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | df0d08693fea0b7045a8d327cacc524e7234729ce71cb33535040e234209b079 |
|
MD5 | 23f4f35b862c9487ef8a2db1143beebf |
|
BLAKE2b-256 | e8482f591900d84a0e1da20ed81c9a8fe94dc614f99e1d499bdcaa353497e57c |