Skip to main content

The Clickhouse plugin for dbt (data build tool)

Project description

clickhouse dbt logo

build

dbt-clickhouse

This plugin ports dbt functionality to Clickhouse.

We do not test over older versions of Clickhouse. The plugin uses syntax that requires version 22.1 or newer.

Installation

Use your favorite Python package manager to install the app from PyPI, e.g.

pip install dbt-clickhouse

Supported features

  • Table materialization
  • View materialization
  • Incremental materialization
  • Seeds
  • Sources
  • Docs generate
  • Tests
  • Snapshots
  • Ephemeral materialization

Usage Notes

Database

The dbt model database.schema.table is not compatible with Clickhouse because Clickhouse does not support a schema. So we use a simple model schema.table, where schema is the Clickhouse's database. Please, don't use default database!

Model Configuration

Option Description Required?
engine The table engine (type of table) to use when creating tables Optional (default: MergeTree())
order_by A tuple of column names or arbitrary expressions. This allows you to create a small sparse index that helps find data faster. Optional (default: tuple())
partition_by A partition is a logical combination of records in a table by a specified criterion. The partition key can be any expression from the table columns. Optional
unique_key A tuple of column names that uniquely identify rows. For more details on uniqueness constraints, see here. Optional
inserts_only This property is relevant only for incremental materialization. If set to True, incremental updates will be inserted directly to the target table without creating intermediate table. This option has the potential of significantly improve performance and avoid memory limitations on big updates. Optional
settings A dictionary with custom settings for INSERT INTO and CREATE AS SELECT queries. Optional

Example Profile

your_profile_name:
  target: dev
  outputs:
    dev:
      type: clickhouse
      schema: [database name]

      # optional
      driver: [http] # http or native.  If not set will autodetermine base one port
      port: [port]  # default 8123
      user: [user] # default 'default'
      host: [db.clickhouse.com] # default localhost
      password: [password] # default ''
      verify: [verify] # default True
      secure: [secure] # default False
      connect_timeout: [10] # default 10 seconds.
      custom_settings: {} # Custom seetings for the connection - default is empty.

Running Tests

This adapter passes all of dbt basic tests as presented in dbt's official docs: https://docs.getdbt.com/docs/contributing/testing-a-new-adapter#testing-your-adapter.

Note: The only feature that is not supported and not tested is Ephemeral materialization.

Tests running command: pytest tests/integration

You can customize a few test params through environment variables. In order to provide custom params you'll need to create test.env file under root (remember not to commit this file!) and define the following env variables inside:

  1. HOST_ENV_VAR_NAME - Default=localhost
  2. USER_ENV_VAR_NAME - your ClickHouse username. Default=default
  3. PASSWORD_ENV_VAR_NAME - your ClickHouse password. Default=''
  4. PORT_ENV_VAR_NAME - ClickHouse client port. Default=8123
  5. RUN_DOCKER_ENV_VAR_NAME - Identify whether to run clickhouse-server docker image (see tests/docker-compose.yml). Default=False. Set it to True if you'd like to raise a docker image (assuming docker-compose is installed in your machine) during tests that launches a clickhouse-server. Note: If you decide to run a docker image you should set PORT_ENV_VAR_NAME to 10900 too.

Original Author

ClickHouse wants to thank @silentsokolov for creating this connector and for their valuable contributions.

Update 05/31/2022

  • Incremental changes of an incremental model are loaded into a MergeTree table instead of in-memory temporary table. This removed memory limitations - Clickhouse recommends that in-memory table engines should not exceed 100 million rows.
  • Incremental model supports 'inserts_only' mode where incremental changes are loaded directly to the target table instead of creating a temporary table for the changes and running another insert-into command. This mode is relevant only for immutable data, and can accelerate dramatically the performance of the incremental materialization.
  • Fix update and delete in snapshots.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

dbt-clickhouse-1.1.6.tar.gz (23.8 kB view details)

Uploaded Source

Built Distribution

dbt_clickhouse-1.1.6-py2.py3-none-any.whl (25.7 kB view details)

Uploaded Python 2 Python 3

File details

Details for the file dbt-clickhouse-1.1.6.tar.gz.

File metadata

  • Download URL: dbt-clickhouse-1.1.6.tar.gz
  • Upload date:
  • Size: 23.8 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/4.0.1 CPython/3.9.13

File hashes

Hashes for dbt-clickhouse-1.1.6.tar.gz
Algorithm Hash digest
SHA256 71f8fcea1387993471b2dc1bc7e29be5ef1c5ee3e97ff734b3563031152c77dd
MD5 cb9d77c264284bbd80b794bdebca9dbe
BLAKE2b-256 05add11583a13a267754975ef7a0780e605f264373c31fe5d4e10e8a7b5aa0ee

See more details on using hashes here.

File details

Details for the file dbt_clickhouse-1.1.6-py2.py3-none-any.whl.

File metadata

File hashes

Hashes for dbt_clickhouse-1.1.6-py2.py3-none-any.whl
Algorithm Hash digest
SHA256 92a61c4a1a981ba8db445a5f7244c9dc6c4137924c9831e41eadc71e64e2afc8
MD5 f22c31108bd8cec8bfb154afb389d1b3
BLAKE2b-256 847c7442a4f1e32d17a15dd602a7c924ca4ddfe085a8650141437a8158e1ccb0

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page