Skip to main content

MCP server for interacting with Keboola Connection

Project description

Keboola MCP Server

CI codecov Keboola Explorer Server MCP server smithery badge

A Model Context Protocol (MCP) server for interacting with Keboola Connection. This server provides tools for listing and accessing data from Keboola Storage API.

Requirements

  • Python 3.10 or newer
  • Keboola Storage API token
  • Snowflake or BigQuery Read Only Workspace

Installation

Installing via Pip

First, create a virtual environment and then install the keboola_mcp_server package:

python3 -m venv --upgrade-deps .venv
source .venv/bin/activate

pip3 install keboola_mcp_server

Installing via Smithery

To install Keboola MCP Server for Claude Desktop automatically via Smithery:

npx -y @smithery/cli install keboola-mcp-server --client claude

Claude Desktop Setup

To use this server with Claude Desktop, follow these steps:

  1. Create or edit the Claude Desktop configuration file:

    • macOS: ~/Library/Application Support/Claude/claude_desktop_config.json
    • Windows: %APPDATA%\Claude\claude_desktop_config.json
  2. Add the following configuration (adjust paths according to your setup):

{
  "mcpServers": {
    "keboola": {
      "command": "uvx",
      "args": [
        "keboola_mcp_server",
        "--api-url",
        "https://connection.YOUR_REGION.keboola.com"
      ],
      "env": {
        "KBC_STORAGE_TOKEN": "your_keboola_storage_token",
        "KBC_WORKSPACE_SCHEMA": "your_workspace_schema"
      }
    }
  }
}

Replace:

  • /path/to/keboola-mcp-server with your actual path to the cloned repository
  • YOUR_REGION with your Keboola region (e.g., north-europe.azure, etc.). You can remove it if your region is just connection explicitly
  • your_keboola_storage_token with your Keboola Storage API token
  • your_workspace_schema with your Snowflake schema or BigQuery dataset of your workspace

Note: If you are using a specific version of Python (e.g. 3.11 due to some package compatibility issues), you'll need to update the command into using that specific version, e.g. /path/to/keboola-mcp-server/.venv/bin/python3.11

Note: The Workspace can be created in your Keboola project. It is the same project where you got your Storage Token. The workspace will provide all the necessary connection parameters including the schema or dataset name.

  1. After updating the configuration:
    • Completely quit Claude Desktop (don't just close the window)
    • Restart Claude Desktop
    • Look for the hammer icon in the bottom right corner, indicating the server is connected

Troubleshooting

If you encounter connection issues:

  1. Check the logs in Claude Desktop for any error messages
  2. Verify your Keboola Storage API token is correct
  3. Ensure all paths in the configuration are absolute paths
  4. Confirm the virtual environment is properly activated and all dependencies are installed

Cursor AI Setup

To use this server with Cursor AI, you have two options for configuring the transport method: Server-Sent Events (SSE) or Standard I/O (stdio).

  1. Create or edit the Cursor AI configuration file:

    • Location: ~/.cursor/mcp.json
  2. Add one of the following configurations (or all) based on your preferred transport method:

Option 1: Using Server-Sent Events (SSE)

{
  "mcpServers": {
    "keboola": {
      "url": "http://localhost:8000/sse?storage_token=YOUR_KEBOOLA_STORAGE_TOKEN&workspace_schema=YOUR_WORKSPACE_SCHEMA"
    }
  }
}

Option 2a: Using Standard I/O (stdio)

{
  "mcpServers": {
    "keboola": {
      "command": "uvx",
      "args": [
        "keboola_mcp_server",
        "--transport",
        "stdio",
         "--api-url",
         "https://connection.YOUR_REGION.keboola.com"
      ],
      "env": {
        "KBC_STORAGE_TOKEN": "your_keboola_storage_token", 
        "KBC_WORKSPACE_SCHEMA": "your_workspace_schema"         
      }
    }
  }
}

Option 2b: Using WSL Standard I/O (wsl stdio)

When running the MCP server from Windows Subsystem for Linux with Cursor AI, use this.

{
  "mcpServers": {
    "keboola": {
      "command": "wsl.exe",
      "args": [
        "bash",
        "-c",
        "'source /wsl_path/to/keboola-mcp-server/.env",
        "&&",
        "/wsl_path/to/keboola-mcp-server/.venv/bin/python -m keboola_mcp_server.cli --transport stdio'"
      ]
    }
  }
}
  • where /wsl_path/to/keboola-mcp-server/.env file contains environment variables:
export KBC_STORAGE_TOKEN="your_keboola_storage_token"
export KBC_WORKSPACE_SCHEMA="your_workspace_schema"

Replace:

  • /path/to/keboola-mcp-server with your actual path to the cloned repository
  • YOUR_REGION with your Keboola region (e.g., north-europe.azure, etc.). You can remove it if your region is just connection explicitly
  • your_keboola_storage_token with your Keboola Storage API token
  • your_workspace_schema with your Snowflake schema or BigQuery dataset of your workspace

After updating the configuration:

  1. Restart Cursor AI
  2. If you use the sse transport make sure to start your MCP server. You can do so by running this in the activated virtual environment where you built the server:
    /path/to/keboola-mcp-server/.venv/bin/python -m keboola_mcp_server --transport sse --api-url https://connection.YOUR_REGION.keboola.com
    
  3. Cursor AI should be automatically detect your MCP server and enable it.

BigQuery support

If your Keboola project uses BigQuery backend you will need to set GOOGLE_APPLICATION_CREDENTIALS environment variable in addition to KBC_STORAGE_TOKEN and KBC_WORKSPACE_SCHEMA.

  1. Go to your Keboola BigQuery workspace and display its credentials (click Connect button).
  2. Download the credentials file to your local disk. It is a plain JSON file.
  3. Set the full path of the downloaded JSON credentials file to GOOGLE_APPLICATION_CREDENTIALS environment variable.

This will give your MCP server instance permissions to access your BigQuery workspace in Google Cloud.

Available Tools

The server provides the following tools for interacting with Keboola Connection:

Storage Tools

  • retrieve_buckets - List all buckets in your Keboola project
  • get_bucket_detail - Get detailed information about a specific bucket
  • retrieve_bucket_tables - List all tables in a specific bucket
  • get_table_detail - Get detailed information about a specific table
  • update_bucket_description - Update the description of a bucket
  • update_table_description - Update the description of a table

SQL Tools

  • query_table - Execute SQL queries on tables in your workspace
  • get_sql_dialect - Get the SQL dialect used in your workspace (Snowflake or BigQuery)

Component Tools

  • retrieve_components - List available components and their configurations
  • retrieve_transformations - List transformation configurations
  • get_component_details - Get detailed information about a specific component
  • create_sql_transformation - Create a new SQL transformation configuration

Job Tools

  • retrieve_jobs - List jobs in your project
  • get_job_detail - Get detailed information about a specific job
  • start_job - Start a new job for a component configuration

Documentation Tools

  • docs_query - Query documentation and help information

Development

Run tests:

pytest

Format code:

black .
isort .

Type checking:

mypy .

License

MIT License - see LICENSE file for details.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distributions

No source distribution files available for this release.See tutorial on generating distribution archives.

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

keboola_mcp_server-0.15.0-py3-none-any.whl (33.6 kB view details)

Uploaded Python 3

File details

Details for the file keboola_mcp_server-0.15.0-py3-none-any.whl.

File metadata

File hashes

Hashes for keboola_mcp_server-0.15.0-py3-none-any.whl
Algorithm Hash digest
SHA256 a21fe6391cdd8f499182d795a44fc26871452041b825a12743bd78974f57d629
MD5 8682d07fc1d3f31b896f5359a2b66567
BLAKE2b-256 b48367243d0f009abc7def9024674ac2fa099df418160b634f4e7b2dba42bf3b

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page