Skip to main content

Commons library for ingesting RDBMS metadata into Google Cloud Data Catalog

Project description

google-datacatalog-rdbms-connector

Common resources for Data Catalog RDBMS connectors.

Python package PyPi License Issues

Disclaimer: This is not an officially supported Google product.

Table of Contents


1. Installation

Install this library in a virtualenv using pip. virtualenv is a tool to create isolated Python environments. The basic problem it addresses is one of dependencies and versions, and indirectly permissions.

With virtualenv, it's possible to install this library without needing system install permissions, and without clashing with the installed system dependencies. Make sure you use Python 3.6+.

1.1. Mac/Linux

pip3 install virtualenv
virtualenv --python python3.6 <your-env>
source <your-env>/bin/activate
<your-env>/bin/pip install google-datacatalog-rdbms-connector

1.2. Windows

pip3 install virtualenv
virtualenv --python python3.6 <your-env>
<your-env>\Scripts\activate
<your-env>\Scripts\pip.exe install google-datacatalog-rdbms-connector

2. Install from source

2.1. Get the code

git clone https://github.com/GoogleCloudPlatform/datacatalog-connectors-rdbms/
cd datacatalog-connectors-rdbms/google-datacatalog-rdbms-connector

2.2. Virtualenv

Using virtualenv is optional, but strongly recommended.

2.2.1. Install Python 3.6
2.2.2. Create and activate a virtualenv
pip3 install virtualenv
virtualenv --python python3.6 <your-env>
source <your-env>/bin/activate
2.2.3. Install
pip install .

3. Developer environment

3.1. Install and run YAPF formatter

pip install --upgrade yapf

# Auto update files
yapf --in-place --recursive src tests

# Show diff
yapf --diff --recursive src tests

# Set up pre-commit hook
# From the root of your git project.
curl -o pre-commit.sh https://raw.githubusercontent.com/google/yapf/master/plugins/pre-commit.sh
chmod a+x pre-commit.sh
mv pre-commit.sh .git/hooks/pre-commit

3.2. Install and run Flake8 linter

pip install --upgrade flake8
flake8 src tests

3.3. Install the package in editable mode (i.e. setuptools “develop mode”)

pip install --editable .

3.4. Run the unit tests

python setup.py test

4. Setting up the RDBMS on a new connector

To set up the RDBMS connector to work with a Relational Database 3 files are needed.

  • metadata_definition.json
  • metadata_query.sql
  • Extending the metadata_scraper class and implementing your rdbms connection method: _create_rdbms_connection

for the metadata_definition file your have fields available for 3 levels:

  • table_container_def
  • table_def
  • column_def

If you want working examples please take a look at the already implemented connectors for: Oracle, Teradata, MySQL, PostgreSQL, Greenplum, Redshift and SQLServer.

For the metadata_defition target fields you have the following options as target:

Level Target Description Mandatory
table_container_def creator Creator of the Table Container. N
table_container_def owner Owner of the Table Container. N
table_container_def update_user Last user that updated the Table Container. N
table_container_def desc Table Container Description. N
table_def num_rows Number of rows contained in the Table. N
table_def creator Creator of the Table. N
table_def owner Owner of the Table. N
table_def update_user Last user that updated the Table. N
table_def desc Table Description. N
table_def table_size_MB Table size, in MB. N

If those fields are configured they will be used to create Tags.

For columns they are used to create the Data Catalog Entry schema, two target fields are used:

Level Target Description Mandatory
column_def type Column type. Y
column_def desc Column description. N

4.1 Add support for optional queries

You can use user configuration file and execute optional queries to scrape additional metadata from a database. You would need to do the following:

  • Create additional SQL queries. At the moment, support for two optional queries are implemented in this common RDBMS package:
    • Statement for refreshing metadata (e.g. ANALYZE)
    • Query to scrape number of rows in each table
  • Extend the query_assembler class and implement a methods for getting optional queries: _get_refresh_statement and _get_path_to_num_rows_query.
  • Implement _get_query_assembler and _execute_refresh_query on the extension of the matadata_scraper class.

You can see working examples of implementing optional queries in the connectors code for PostgreSQL and MySQL. Please also refer to these samples to see how user configuration file ingest_cfg.yaml should look like.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

Built Distribution

google_datacatalog_rdbms_connector-0.8.0-py2.py3-none-any.whl (25.4 kB view details)

Uploaded Python 2 Python 3

File details

Details for the file google-datacatalog-rdbms-connector-0.8.0.tar.gz.

File metadata

  • Download URL: google-datacatalog-rdbms-connector-0.8.0.tar.gz
  • Upload date:
  • Size: 17.9 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.2.0 pkginfo/1.5.0.1 requests/2.24.0 setuptools/40.8.0 requests-toolbelt/0.9.1 tqdm/4.49.0 CPython/3.7.0

File hashes

Hashes for google-datacatalog-rdbms-connector-0.8.0.tar.gz
Algorithm Hash digest
SHA256 13a3d957ed9cac8aaca7c185f4c6b5e6c1ceac889626d4f0be4b8af61c2782db
MD5 dcec853c9c9d84af57312a7fbdfc764c
BLAKE2b-256 faa41f8d7fd8c8cb30c04f1b0dd64381620bfdd1123e1bad846e0d9453a2b567

See more details on using hashes here.

File details

Details for the file google_datacatalog_rdbms_connector-0.8.0-py2.py3-none-any.whl.

File metadata

File hashes

Hashes for google_datacatalog_rdbms_connector-0.8.0-py2.py3-none-any.whl
Algorithm Hash digest
SHA256 22a0502366d375eaf0deaddb77e5200229ba846837a2b5ff0f7623c6367fd27e
MD5 eae0a1bf1cd9f7c8b373fca40e0b76de
BLAKE2b-256 c6f4dab049b349f07991be64cb5d166acd5cdf7666de8d8189d80a616be173dd

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page