Skip to main content

A simple writing utility for writing to QuestDB from PySpark / AWS Glue using Line Protocol and Sockets

Project description

AWS Glue / PySpark QuestDB writer

PyPI version PyPI - Downloads PyPI - Python Version GitHub issues GitHub forks GitHub stars GitHub license

A simple writing utility for writing to QuestDB from PySpark / AWS Glue

#Introduction A very simple convenience library created due to difficulty in getting custom builds of the AWS Glue Libs for AWS for local development. The standard release is Python 3.6 and requires some changes to be made in order to add in extra libraries.

The InfluxDB writer is a potential alternative to this, but I didn't have much of a chance to get it working due to dependencies and it not easily supporting PySpark.

#Installation

Install this via pip

pip3 install awsglue-questdb-writer

#Usage

In your AWS Glue / PySpark Job include the file via

from awsglue_questdb_writer import *

Usage is by passing a DF to the function, this should ideally be a DF from a SparkSQL output like Glue creates (e.g. from the from_catalog) as that is what has been tested.

Important to note:

  • All Timestamps must be datetime objects
  • Nanosecond precision (required by QuestDB) is currently only your timestamp precision with added zeros
  • If you need real nanosecond permission you must be on Python 3.7 and update the library to use it (See comments)
  • QuestDB is whitespace sensitive, all datetimes are quoted but any other fields with whitespace will cause this to fail (silently)
  • There is no socket response from this library (it is designed to be unmonitored and high throughput) - if errors are in your input it will fail silently (PR's welcome)
  • There is a convenience line to drop unwanted fields prior to passing this into the function to write to QuestDB
args = getResolvedOptions(sys.argv,
                          ['TempDir', 'JOB_NAME', 'db_name', 'temp_workflow_bucket', 'questdb_host', 'questdb_port'])

sc = SparkContext()
glueContext = GlueContext(sc)
spark = glueContext.spark_session
job = Job(glueContext)
job.init(args['JOB_NAME'], args)

allDaily = glueContext.create_dynamic_frame.from_catalog(database=args['db_name'],
                                                         table_name="daily",
                                                         transformation_ctx="allDaily",
                                                         )

df = allDaily.toDF()
tdf = df.withColumn('reading_date_time', F.to_timestamp(df['reading_date_time'], '%Y-%m-%dT%H:%M:%S.%f'))
tdf = tdf.drop(*["ingestion_date", "period_start", "period_end", "quality_method",
                 "event", "import_reactive_total", "export_reactive_total"])

write_to_quest(df=tdf, measurement="meter_id", table="daily", timestamp_field="reading_date_time", args=args)

job.commit()

#License See LICENSE for full details

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Files for awsglue-questdb-writer, version 0.0.2
Filename, size File type Python version Upload date Hashes
Filename, size awsglue_questdb_writer-0.0.2-py3-none-any.whl (5.1 kB) File type Wheel Python version py3 Upload date Hashes View
Filename, size awsglue-questdb-writer-0.0.2.tar.gz (4.8 kB) File type Source Python version None Upload date Hashes View

Supported by

AWS AWS Cloud computing Datadog Datadog Monitoring DigiCert DigiCert EV certificate Facebook / Instagram Facebook / Instagram PSF Sponsor Fastly Fastly CDN Google Google Object Storage and Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Salesforce Salesforce PSF Sponsor Sentry Sentry Error logging StatusPage StatusPage Status page