Skip to main content

Package for ingesting Vertica metadata into Google Cloud Data Catalog

Project description

google-datacatalog-vertica-connector

Package for ingesting Vertica metadata into Google Cloud Data Catalog.

Python package PyPi License Issues

Disclaimer: This is not an officially supported Google product.

Table of Contents


1. Installation

Install this library in a virtualenv using pip. virtualenv is a tool to create isolated Python environments. The basic problem it addresses is one of dependencies and versions, and indirectly permissions.

With virtualenv, it's possible to install this library without needing system install permissions, and without clashing with the installed system dependencies. Make sure you use Python 3.6+.

1.1. Mac/Linux

pip3 install virtualenv
virtualenv --python python3.6 <your-env>
source <your-env>/bin/activate
<your-env>/bin/pip install google-datacatalog-vertica-connector

1.2. Windows

pip3 install virtualenv
virtualenv --python python3.6 <your-env>
<your-env>\Scripts\activate
<your-env>\Scripts\pip.exe install google-datacatalog-vertica-connector

1.3. Install from source

1.3.1. Get the code

git clone https://github.com/GoogleCloudPlatform/datacatalog-connectors-rdbms/
cd datacatalog-connectors-rdbms/google-datacatalog-vertica-connector

1.3.2. Create and activate a virtualenv

pip3 install virtualenv
virtualenv --python python3.6 <your-env>
source <your-env>/bin/activate

1.3.3. Install the library

pip install .

2. Environment setup

2.1. Auth credentials

2.1.1. Create a service account and grant it below roles

  • Data Catalog Admin

2.1.2. Download a JSON key and save it as

  • <YOUR-CREDENTIALS_FILES_FOLDER>/vertica2dc-datacatalog-credentials.json

Please notice this folder and file will be required in next steps.

2.2. Set environment variables

Replace below values according to your environment:

export GOOGLE_APPLICATION_CREDENTIALS=data_catalog_credentials_file

3. Run entry point

3.1. Run Python entry point

  • Virtualenv

Connect to a server:

vertica2datacatalog \
  --vertica-host <VERTICA-HOST-IP> \
  --vertica-user <VERTICA-USER> \
  --vertica-pass <VERTICA-PASSWORD> \
  --datacatalog-project-id <YOUR-DATACATALOG-PROJECT-ID> \
  --datacatalog-location-id <YOUR-DATACATALOG-LOCATION-ID>

Read metadata from a CSV file:

vertica2datacatalog \
  --raw-metadata-csv <PATH-TO-A-FULL-DUMP-CSV-FILE> \
  --datacatalog-project-id <YOUR-DATACATALOG-PROJECT-ID> \
  --datacatalog-location-id <YOUR-DATACATALOG-LOCATION-ID>

3.2. Run Docker entry point

docker build --rm --tag vertica2datacatalog .

Connect to a server:

docker run --rm --tty -v <YOUR-CREDENTIALS_FILES_FOLDER>:/data \
  vertica2datacatalog \ 
  --vertica-host <VERTICA-HOST-IP> \
  --vertica-user <VERTICA-USER> \
  --vertica-pass <VERTICA-PASSWORD> \
  --datacatalog-project-id <YOUR-DATACATALOG-PROJECT-ID> \
  --datacatalog-location-id <YOUR-DATACATALOG-LOCATION-ID>

Read metadata from a CSV file:

docker run --rm --tty -v <YOUR-CREDENTIALS_FILES_FOLDER>:/data \
  vertica2datacatalog \ 
  --raw-metadata-csv /data/<PATH-TO-A-FULL-DUMP-CSV-FILE> \
  --datacatalog-project-id <YOUR-DATACATALOG-PROJECT-ID> \
  --datacatalog-location-id <YOUR-DATACATALOG-LOCATION-ID>

4. Developer environment

4.1. Install and run Yapf formatter

pip install --upgrade yapf

# Auto update files
yapf --in-place --recursive src tests

# Show diff
yapf --diff --recursive src tests

# Set up pre-commit hook
# From the root of your git project.
curl -o pre-commit.sh https://raw.githubusercontent.com/google/yapf/master/plugins/pre-commit.sh
chmod a+x pre-commit.sh
mv pre-commit.sh .git/hooks/pre-commit

4.2. Install and run Flake8 linter

pip install --upgrade flake8
flake8 src tests

4.3. Run Tests

python setup.py test

5. Troubleshooting

In the case a connector execution hits Data Catalog quota limit, an error will be raised and logged with the following detailement, depending on the performed operation READ/WRITE/SEARCH:

status = StatusCode.RESOURCE_EXHAUSTED
details = "Quota exceeded for quota metric 'Read requests' and limit 'Read requests per minute' of service 'datacatalog.googleapis.com' for consumer 'project_number:1111111111111'."
debug_error_string = 
"{"created":"@1587396969.506556000", "description":"Error received from peer ipv4:172.217.29.42:443","file":"src/core/lib/surface/call.cc","file_line":1056,"grpc_message":"Quota exceeded for quota metric 'Read requests' and limit 'Read requests per minute' of service 'datacatalog.googleapis.com' for consumer 'project_number:1111111111111'.","grpc_status":8}"

For more info about Data Catalog quota, go to: Data Catalog quota docs.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

Built Distribution

File details

Details for the file google-datacatalog-vertica-connector-0.9.0.tar.gz.

File metadata

  • Download URL: google-datacatalog-vertica-connector-0.9.0.tar.gz
  • Upload date:
  • Size: 9.8 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.4.1 importlib_metadata/3.7.3 pkginfo/1.7.0 requests/2.25.1 requests-toolbelt/0.9.1 tqdm/4.59.0 CPython/3.8.8

File hashes

Hashes for google-datacatalog-vertica-connector-0.9.0.tar.gz
Algorithm Hash digest
SHA256 77f37ddf49b1f5ffd8a347e3803a1bb9c0cc6160e4f14835daf03509b0ef2f99
MD5 1a24bf63cc8d2eb8421356b218f67a0f
BLAKE2b-256 9f11d7b07f69b8bac16ecc4ab87812d37829fdf77e9a7f1f5ea5084d9ba587ed

See more details on using hashes here.

File details

Details for the file google_datacatalog_vertica_connector-0.9.0-py2.py3-none-any.whl.

File metadata

File hashes

Hashes for google_datacatalog_vertica_connector-0.9.0-py2.py3-none-any.whl
Algorithm Hash digest
SHA256 a97b28af1585d101c14e557be965a606d88546eae61215f89e3a9e7083fd5cdf
MD5 db8230fb8c934d1836b2079f68a3d9ad
BLAKE2b-256 fe11420b4624bed0e3e17ba3b2e7fc98a04fe824c003558126019a9ac8cf45dd

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page