Skip to main content

A simple tool to compile and run CUDA kernels on local GPUs or Modal cloud infrastructure, now with MCP server support

Project description

minimodal

A simple command-line tool to compile and run CUDA kernels on local GPUs or Modal cloud infrastructure, now with MCP (Model Context Protocol) server support.

Features:

  • Automatic GPU detection and architecture mapping
  • Seamless fallback to Modal cloud GPUs when local GPU isn't available
  • Support for include directories and custom nvcc flags
  • Works with Modal GPU naming convention (A10G, H100, B200, etc.)
  • MCP server enabling AI agents to compile and run CUDA code

Installation

uv pip install minimodal

Or install from source:

cd minimodal
uv pip install -e .

Usage

Basic usage (defaults to A10G):

minimodal v1.cu

With a specific GPU:

minimodal v1.cu --gpu H100

With include directories:

minimodal v1.cu --gpu B200 -I ./includes

With execution arguments:

minimodal v1.cu --gpu A100 --exec-args "--N 2097152 --warmup 5"

With custom nvcc flags:

minimodal kernel.cu --gpu H100 -I ./includes --nvcc-flags "-O3 --ptxas-options=-v"

Supported GPUs

The tool automatically detects GPU architecture for common GPU models:

  • H100 / H800 → sm_90
  • A100 / A800 → sm_80
  • A10G → sm_86
  • B200 / B100 → sm_100a
  • RTX4090 → sm_89
  • RTX3090 / RTX3080 → sm_86
  • V100 → sm_70
  • T4 → sm_75

You can also specify the architecture directly using --arch:

minimodal kernel.cu --arch 100a

Options

  • --gpu, -g: GPU name (e.g., A10G, H100, B200, A100) or architecture (e.g., 100a, 90). Default: A10G
  • --include-dir, -I: Include directory. Can be specified multiple times.
  • --exec-args: Arguments to pass to the compiled CUDA executable.
  • --arch: Override architecture. If not specified, inferred from --gpu.
  • --output, -o: Output executable path. If not specified, uses temporary file.
  • --keep-binary: Keep the compiled binary after execution.
  • --nvcc-flags: Additional flags to pass to nvcc (e.g., '-O3 --ptxas-options=-v').

MCP Server Usage

The minimodal package includes an MCP (Model Context Protocol) server that enables AI agents to compile and run CUDA kernels. This allows you to use minimodal with tools like Claude Desktop, Cursor, or other MCP-compatible AI assistants.

Starting the MCP Server

Start the server with stdio transport (default):

minimodal-mcp

Start the server with SSE transport (for web clients):

minimodal-mcp --transport sse

Start with SSE on a custom port:

minimodal-mcp --transport sse --port 9000

Configuring Claude Desktop

  1. Add the following configuration to your Claude Desktop config file (~/Library/Application Support/Claude/claude_desktop_config.json on macOS):
{
  "mcpServers": {
    "minimodal": {
      "command": "minimodal-mcp",
      "args": []
    }
  }
}
  1. Restart Claude Desktop

  2. The minimodal tools will now be available in your Claude conversations

Available MCP Tools

The MCP server exposes the following tools:

  • compile_cuda: Compile and run a CUDA kernel file
  • list_gpus: List available local and Modal cloud GPUs
  • get_gpu_architecture: Get CUDA architecture for a specific GPU name

Example MCP Usage

Once configured with an AI assistant, you can use natural language to compile CUDA code:

  • "Compile and run the vector_add.cu file on an H100 GPU"
  • "Check what GPUs are available"
  • "What architecture does an A10G GPU use?"
  • "Compile kernel.cu with include directories include/, optimization flags, and keep the binary"

Requirements

  • CUDA toolkit with nvcc compiler
  • NVIDIA GPU drivers
  • nvidia-smi command-line tool
  • Modal account and API key (for cloud GPU usage)

Examples

Compile and run a CUDA kernel (defaults to A10G):

minimodal v1.cu

Compile and run on a specific GPU:

minimodal v1.cu --gpu H100

Compile with multiple include directories:

minimodal kernel.cu --gpu B200 -I ./includes -I ./headers

Run with custom arguments:

minimodal benchmark.cu --gpu A100 --exec-args "--size 4096 --iterations 100"

Development

To develop locally:

git clone <repository-url>
cd minimodal
uv pip install -e .

Publishing to PyPI

# Build distribution
uv build

# Upload to PyPI
uv pip install twine
twine upload dist/*

License

MIT

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

minimodal-0.2.0.tar.gz (13.5 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

minimodal-0.2.0-py3-none-any.whl (14.9 kB view details)

Uploaded Python 3

File details

Details for the file minimodal-0.2.0.tar.gz.

File metadata

  • Download URL: minimodal-0.2.0.tar.gz
  • Upload date:
  • Size: 13.5 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.12

File hashes

Hashes for minimodal-0.2.0.tar.gz
Algorithm Hash digest
SHA256 4620b50e76af2ad84fe603c3724edf61fd9e4ba9718a04d6ede452fa37beeb29
MD5 30679091fadc84a3c3b697d690c4bc5e
BLAKE2b-256 3aaf9918a8e99282416de5cc4f9fa5f9f4993fde99809f80cf35b5dcd895d5a2

See more details on using hashes here.

File details

Details for the file minimodal-0.2.0-py3-none-any.whl.

File metadata

  • Download URL: minimodal-0.2.0-py3-none-any.whl
  • Upload date:
  • Size: 14.9 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.12

File hashes

Hashes for minimodal-0.2.0-py3-none-any.whl
Algorithm Hash digest
SHA256 cdd78e12b6bf9d622393bfc19692daf50bef51d20efa7b430527460f5ef351f2
MD5 84859744af0623e05001ed9ea9501a55
BLAKE2b-256 175e1fc6a8869789079dfc9e245e13c142f6eae02dbbe38fd47db2b2fa4c7aff

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page