MCPStack tool exposing the AutoDDG dataset description workflow
Project description
[!IMPORTANT] If you haven’t visited the MCPStack main orchestrator repository yet, please start there: MCPStack
[!CAUTION] Please be aware that this MCP is in an early-alpha stage. While it is functional and can be used for various tasks, it may still contain bugs or incomplete features. Feel free to report any issues you encounter or suggest improvements. Even better, feel free to contribute directly!
[!WARNING] Please be aware that you cannot use this MCP without an OpenAI-compatible API key. To gen. one, please visit: https://platform.openai.com/account/api-keys
[!NOTE] For the time being, this MCP is working with the branch
feat/modern_pythonic_library_transformationfrom the mother library, AutoDDG. See more at: https://github.com/VIDA-NYU/AutoDDG/pull/4. As such, we recommend you to install AutoDDG from source with this library, until the PR is merged upstream. Adapt the "autoddg" in the dependencies inpyproject.tomlaccordingly.
💡 About The MCPStack AutoDDG Tool
This repository provides a native MCP around the AutoDDG library for dataset description and discovery:
- Load a CSV and keep a deterministic sample (by size or percentage).
- Profile a dataframe (datamart-style notes).
- Infer a semantic profile for columns.
- Generate a concise topic.
- Produce a readable dataset description.
- Expand that description for search/discovery (tune the
temperatureetc.). - Optionally evaluate the description with a separate evaluator key.
AutoDDG official library (without the MCP wrapper): https://github.com/VIDA-NYU/AutoDDG
Installation
The tool is distributed as a standard Python package. MCPStack will auto-discover it.
Via uv (recommended)
uv add mcpstack-autoddg
Via pip
pip install mcpstack-autoddg
(Dev) Pre-commit hooks
uv run pre-commit install
# or: pre-commit install
Using With MCPStack — CLI workflow
This tool declares entry points so MCPStack can see it automatically:
[project.entry-points."mcpstack.tools"]
autoddgtool = "mcpstack_autoddg.tool:AutoDDGTool"
[project.entry-points."mcpstack.tool_clis"]
autoddgtool = "mcpstack_autoddg.cli:AutoDDGCLI.get_app"
1) (Optional) Configure environment
AutoDDG requires an OpenAI-compatible key. You may optionally provide a separate evaluator key:
AUTO_DDG_OPENAI_API_KEY: "<your key>" (required for generation)
AUTO_DDG_EVALUATOR_API_KEY: "<your key>" (optional; falls back to AUTO_DDG_OPENAI_API_KEY)
Use the CLI to generate a config file (useful for CI or sharing defaults):
uv run mcpstack tools autoddg configure
# Then is followed an interactive prompt to config and set parameters.
Or you can pass parameters directly, e.g.:
uv run mcpstack tools autoddg configure \
--model-name gpt-4o-mini \
--description-words 120 \
--description-temperature 0.0 \
--topic-temperature 0.0 \
--api-key sk-... \
--evaluator-key sk-... \
-o autoddg_config.json \
--verbose
For others, feel free to uv run mcpstack tools autoddg --help to see all options.
2) Add to a pipeline
Create or extend a pipeline with AutoDDG:
# New pipeline
uv run mcpstack pipeline autoddg --new-pipeline my_pipeline.json --tool-config autoddg_config.json
# Or append to an existing one
uv run mcpstack pipeline autoddg --to-pipeline my_pipeline.json --tool-config autoddg_config.json
Programmatic API Workflow
Use the AutoDDG tool directly in a stack:
from MCPStack.stack import MCPStackCore
from mcpstack_autoddg import AutoDDGTool
pipeline = (
MCPStackCore()
.with_tool(AutoDDGTool(
model_name="gpt-4o-mini",
search_model_name=None,
semantic_model_name=None,
description_words=120,
description_temperature=0.0,
topic_temperature=0.0,
evaluator_model_name="gpt-4o",
))
.build(type="fastmcp", save_path="autoddg_pipeline.json")
.run()
)
AutoDDG Actions Supported
[!NOTE] If any action fails, feel free to open an issue so we may update with the potential changes on the mother library, AutoDDG. https://github.com/VIDA-NYU/AutoDDG
load_dataset(csv_path|csv_text, sample_size?, sample_percent?, random_state=9)→ load CSV and store a sampled CSV string in stateprofile_dataset()→ datamart-like profile; may also return semantic notesgenerate_semantic_profile()→ infer semantic metadata for columnsgenerate_topic(title, original_description?, dataset_sample?)→ concise dataset topicgenerate_description(dataset_sample?, use_profile=True, use_semantic_profile=True, use_topic=True)→ readable description; enforces prerequisites if the flags are left onexpand_description_for_search()→ search-oriented variant of the last description (needs a topic)evaluate_description()→ runs evaluator (requires evaluator key or reuse of generation key)get_state_summary()→ booleans for which artifacts exist in state
License
MIT — see LICENSE.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file mcpstack_autoddg-0.0.0.tar.gz.
File metadata
- Download URL: mcpstack_autoddg-0.0.0.tar.gz
- Upload date:
- Size: 12.4 kB
- Tags: Source
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
7c9d00cd087770e916923cd4407b042815902ab178c1fcdfc180b6e8cafb7332
|
|
| MD5 |
bc48d9afad1d05f4bf71ce3c3a3e4765
|
|
| BLAKE2b-256 |
863bdcd0b6a328f5f854e5c65ddff9b283155529694152135cb62802f6902495
|
Provenance
The following attestation bundles were made for mcpstack_autoddg-0.0.0.tar.gz:
Publisher:
publish.yaml on MCP-Pipeline/mcpstack-autoddg
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
mcpstack_autoddg-0.0.0.tar.gz -
Subject digest:
7c9d00cd087770e916923cd4407b042815902ab178c1fcdfc180b6e8cafb7332 - Sigstore transparency entry: 537990612
- Sigstore integration time:
-
Permalink:
MCP-Pipeline/mcpstack-autoddg@ba7136cbb95a6129fb1ba0aa2496ae1d1f2b93cc -
Branch / Tag:
refs/tags/v0.0.1 - Owner: https://github.com/MCP-Pipeline
-
Access:
private
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
publish.yaml@ba7136cbb95a6129fb1ba0aa2496ae1d1f2b93cc -
Trigger Event:
push
-
Statement type:
File details
Details for the file mcpstack_autoddg-0.0.0-py3-none-any.whl.
File metadata
- Download URL: mcpstack_autoddg-0.0.0-py3-none-any.whl
- Upload date:
- Size: 12.1 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
a850f16b75629883f97692c0136f2c2320fe346715340bdea6bb34eb17494ae8
|
|
| MD5 |
25aadddefe11d964d46a42760a1cfda7
|
|
| BLAKE2b-256 |
62d375434ed477a644d9bdfcf5aea544d7106c584ffacff22eb65c8fa0e050b2
|
Provenance
The following attestation bundles were made for mcpstack_autoddg-0.0.0-py3-none-any.whl:
Publisher:
publish.yaml on MCP-Pipeline/mcpstack-autoddg
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
mcpstack_autoddg-0.0.0-py3-none-any.whl -
Subject digest:
a850f16b75629883f97692c0136f2c2320fe346715340bdea6bb34eb17494ae8 - Sigstore transparency entry: 537990628
- Sigstore integration time:
-
Permalink:
MCP-Pipeline/mcpstack-autoddg@ba7136cbb95a6129fb1ba0aa2496ae1d1f2b93cc -
Branch / Tag:
refs/tags/v0.0.1 - Owner: https://github.com/MCP-Pipeline
-
Access:
private
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
publish.yaml@ba7136cbb95a6129fb1ba0aa2496ae1d1f2b93cc -
Trigger Event:
push
-
Statement type: