Skip to main content

Singer.io tap for extracting data from PostgresSQL - PipelineWise compatible

Project description

pipelinewise-tap-postgres

PyPI version PyPI - Python Version License: MIT

Singer tap that extracts data from a PostgreSQL database and produces JSON-formatted data following the Singer spec.

This is a PipelineWise compatible tap connector.

How to use it

The recommended method of running this tap is to use it from PipelineWise. When running it from PipelineWise you don't need to configure this tap with JSON files and most of things are automated. Please check the related documentation at Tap Postgres

If you want to run this Singer Tap independently please read further.

Install and Run

First, make sure Python 3 is installed on your system or follow these installation instructions for Mac or Ubuntu.

It's recommended to use a virtualenv:

  python3 -m venv venv
  pip install pipelinewise-tap-postgres

or

  make venv

Create a config.json

{
  "host": "localhost",
  "port": 5432,
  "user": "postgres",
  "password": "secret",
  "dbname": "db"
}

These are the same basic configuration properties used by the PostgreSQL command-line client (psql).

Full list of options in config.json:

Property Type Required? Default Description
host String Yes - PostgreSQL host
port Integer Yes - PostgreSQL port
user String Yes - PostgreSQL user
password String Yes - PostgreSQL password
dbname String Yes - PostgreSQL database name
filter_schemas String No None Comma separated schema names to scan only the required schemas to improve the performance of data extraction.
ssl String No None If set to "true" then use SSL via postgres sslmode require option. If the server does not accept SSL connections or the client certificate is not recognized the connection will fail.
logical_poll_total_seconds Integer No 10800 Stop running the tap when no data received from wal after certain number of seconds.
break_at_end_lsn Boolean No true Stop running the tap if the newly received lsn is after the max lsn that was detected when the tap started.
max_run_seconds Integer No 43200 Stop running the tap after certain number of seconds.
debug_lsn String No None If set to "true" then add _sdc_lsn property to the singer messages to debug postgres LSN position in the WAL stream.
tap_id String No None ID of the pipeline/tap
itersize Integer No 20000 Size of PG cursor iterator when doing INCREMENTAL or FULL_TABLE
default_replication_method String No None Default replication method to use when no one is provided in the catalog (Values: LOG_BASED, INCREMENTAL or FULL_TABLE)
use_secondary Boolean No False Use a database replica for INCREMENTAL and FULL_TABLE replication
secondary_host String No - PostgreSQL Replica host (required if use_secondary is True)
secondary_port Integer No - PostgreSQL Replica port (required if use_secondary is True)
limit Integer No None Adds a limit to INCREMENTAL queries to limit the number of records returns per run

Run the tap in Discovery Mode

tap-postgres --config config.json --discover                # Should dump a Catalog to stdout
tap-postgres --config config.json --discover > catalog.json # Capture the Catalog

Add Metadata to the Catalog

Each entry under the Catalog's "stream" key will need the following metadata:

{
  "streams": [
    {
      "stream_name": "my_topic"
      "metadata": [{
        "breadcrumb": [],
        "metadata": {
          "selected": true,
          "replication-method": "LOG_BASED",
        }
      }]
    }
  ]
}

The replication method can be one of FULL_TABLE, INCREMENTAL or LOG_BASED.

Note: Log based replication requires a few adjustments in the source postgres database, please read further for more information.

Run the tap in Sync Mode

tap-postgres --config config.json --catalog catalog.json

The tap will write bookmarks to stdout which can be captured and passed as an optional --state state.json parameter to the tap for the next sync.

Log Based replication requirements

  • PostgreSQL databases running PostgreSQL versions 9.4.x or greater. To avoid a critical PostgreSQL bug, use at least one of the following minor versions:

    • PostgreSQL 12.0
    • PostgreSQL 11.2
    • PostgreSQL 10.7
    • PostgreSQL 9.6.12
    • PostgreSQL 9.5.16
    • PostgreSQL 9.4.21
  • A connection to the master instance. Log-based replication will only work by connecting to the master instance.

  • wal2json plugin: To use Log Based for your PostgreSQL integration, you must install the wal2json plugin version >= 2.3. The wal2json plugin outputs JSON objects for logical decoding, which the tap then uses to perform Log-based Replication. Steps for installing the plugin vary depending on your operating system. Instructions for each operating system type are in the wal2json’s GitHub repository:

  • postgres config file: Locate the database configuration file (usually postgresql.conf) and define the parameters as follows:

    wal_level=logical
    max_replication_slots=5
    max_wal_senders=5
    

    Restart your PostgreSQL service to ensure the changes take effect.

    Note: For max_replication_slots and max_wal_senders, we’re defaulting to a value of 5. This should be sufficient unless you have a large number of read replicas connected to the master instance.

  • Existing replication slot: Log based replication requires a dedicated logical replication slot. In PostgreSQL, a logical replication slot represents a stream of database changes that can then be replayed to a client in the order they were made on the original server. Each slot streams a sequence of changes from a single database.

    Login to the master instance as a superuser and using the wal2json plugin, create a logical replication slot:

      SELECT *
      FROM pg_create_logical_replication_slot('pipelinewise_<database_name>', 'wal2json');
    

    Note: Replication slots are specific to a given database in a cluster. If you want to connect multiple databases - whether in one integration or several - you’ll need to create a replication slot for each database.

To run tests:

  1. Install python test dependencies in a virtual env:
 make venv
  1. You need to have a postgres database to run the tests and export its credentials.

You can make use of the local docker-compose to spin up a test database by running make start_db

Test objects will be created in the postgres database.

  1. To run the unit tests:
  make unit_test
  1. To run the integration tests:
  make integration_test

To run pylint:

Install python dependencies and run python linter

  make venv
  make pylint

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

pipelinewise-tap-postgres-2.1.0.tar.gz (36.8 kB view details)

Uploaded Source

Built Distribution

pipelinewise_tap_postgres-2.1.0-py3-none-any.whl (41.9 kB view details)

Uploaded Python 3

File details

Details for the file pipelinewise-tap-postgres-2.1.0.tar.gz.

File metadata

File hashes

Hashes for pipelinewise-tap-postgres-2.1.0.tar.gz
Algorithm Hash digest
SHA256 aa0b2cd7bb7c897497a28251a39657f67452a2ea90bfa21aaf1a73f565679c50
MD5 89866c6949674be0b1dbd388dc617b3e
BLAKE2b-256 c2bfb8924b58bfefca08a5b1639aea1a79355dc8bb9b3e4d0707730edc143b3f

See more details on using hashes here.

File details

Details for the file pipelinewise_tap_postgres-2.1.0-py3-none-any.whl.

File metadata

File hashes

Hashes for pipelinewise_tap_postgres-2.1.0-py3-none-any.whl
Algorithm Hash digest
SHA256 8e06345ad6056a640f79b6cf0663705e50d76a15f640142a6b20b24565d133e7
MD5 2aeb5fc3c01728b6bd5f4a58d51ae189
BLAKE2b-256 1998fa3bbfae78f95aacb9cc955c384113361534ef022b6d98b90876926f6592

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page