A simple writing utility for writing to QuestDB from PySpark / AWS Glue
Project description
AWS Glue / PySpark QuestDB writer
A simple writing utility for writing to QuestDB from PySpark / AWS Glue
#Introduction A very simple convenience library created due to difficulty in getting custom builds of the AWS Glue Libs for AWS for local development. The standard release is Python 3.6 and requires some changes to be made in order to add in extra libraries.
The InfluxDB writer is a potential alternative to this, but I didn't have much of a chance to get it working due to dependencies and it not easily supporting PySpark.
#Installation
Install this via pip
pip3 install awsglue-questdb-writer
#Usage
In your AWS Glue / PySpark Job include the file via
from awsglue_questdb_writer import *
Usage is by passing a DF to the function, this should ideally be a DF from a SparkSQL output like Glue creates (e.g. from the from_catalog) as that is what has been tested.
Important to note:
- All Timestamps must be datetime objects
- Nanosecond precision (required by QuestDB) is currently only your timestamp precision with added zeros
- If you need real nanosecond permission you must be on Python 3.7 and update the library to use it (See comments)
- QuestDB is whitespace sensitive, all datetimes are quoted but any other fields with whitespace will cause this to fail (silently)
- There is no socket response from this library (it is designed to be unmonitored and high throughput) - if errors are in your input it will fail silently (PR's welcome)
- There is a convenience line to drop unwanted fields prior to passing this into the function to write to QuestDB
args = getResolvedOptions(sys.argv,
['TempDir', 'JOB_NAME', 'db_name', 'temp_workflow_bucket', 'questdb_host', 'questdb_port'])
sc = SparkContext()
glueContext = GlueContext(sc)
spark = glueContext.spark_session
job = Job(glueContext)
job.init(args['JOB_NAME'], args)
allDaily = glueContext.create_dynamic_frame.from_catalog(database=args['db_name'],
table_name="daily",
transformation_ctx="allDaily",
)
df = allDaily.toDF()
tdf = df.withColumn('reading_date_time', F.to_timestamp(df['reading_date_time'], '%Y-%m-%dT%H:%M:%S.%f'))
tdf = tdf.drop(*["ingestion_date", "period_start", "period_end", "quality_method",
"event", "import_reactive_total", "export_reactive_total"])
write_to_quest(df=tdf, measurement="meter_id", table="daily", timestamp_field="reading_date_time", args=args)
job.commit()
#License See LICENSE for full details
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file awsglue-questdb-writer-0.0.1.tar.gz
.
File metadata
- Download URL: awsglue-questdb-writer-0.0.1.tar.gz
- Upload date:
- Size: 4.7 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/3.4.2 importlib_metadata/4.6.1 pkginfo/1.7.1 requests/2.22.0 requests-toolbelt/0.9.1 tqdm/4.61.2 CPython/3.8.10
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | bb8d558640cf2480b499da854f9c4b637345330546233fc18e303ca9f0b8a2aa |
|
MD5 | dccd0ac8c3a255be452ea97f9d601a1b |
|
BLAKE2b-256 | 9ac5ff7e3a94fd15a5cd10f8b250773328e42eb9ce62b8fd88fbf15bc7eae64b |
File details
Details for the file awsglue_questdb_writer-0.0.1-py3-none-any.whl
.
File metadata
- Download URL: awsglue_questdb_writer-0.0.1-py3-none-any.whl
- Upload date:
- Size: 5.0 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/3.4.2 importlib_metadata/4.6.1 pkginfo/1.7.1 requests/2.22.0 requests-toolbelt/0.9.1 tqdm/4.61.2 CPython/3.8.10
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | dd90911aae174138297f346876422cc777c9dfd7ae996e131bfc261f2f6ad452 |
|
MD5 | a2f83cf5aee0acdd19839849f6eec817 |
|
BLAKE2b-256 | 4438344618b395c567fcfedf50087fb0e727d952c6f2f5288f27855a1bfc60af |