Skip to main content

A simple MCP server that provides a unified interface to various LLM providers using Pydantic AI

Project description

LLM Bridge MCP

LLM Bridge MCP allows AI agents to interact with multiple large language models through a standardized interface. It leverages the Message Control Protocol (MCP) to provide seamless access to different LLM providers, making it easy to switch between models or use multiple models in the same application.

Features

  • Unified interface to multiple LLM providers:
    • OpenAI (GPT models)
    • Anthropic (Claude models)
    • Google (Gemini models)
    • DeepSeek
    • ...
  • Built with Pydantic AI for type safety and validation
  • Supports customizable parameters like temperature and max tokens
  • Provides usage tracking and metrics

Tools

The server implements the following tool:

run_llm(
    prompt: str,
    model_name: KnownModelName = "openai:gpt-4o-mini",
    temperature: float = 0.7,
    max_tokens: int = 8192,
    system_prompt: str = "",
) -> LLMResponse
  • prompt: The text prompt to send to the LLM
  • model_name: Specific model to use (default: "openai:gpt-4o-mini")
  • temperature: Controls randomness (0.0 to 1.0)
  • max_tokens: Maximum number of tokens to generate
  • system_prompt: Optional system prompt to guide the model's behavior

Installation

  1. Clone the repository:
git clone https://github.com/yourusername/llm-bridge-mcp.git
cd llm-bridge-mcp
  1. Install uv (if not already installed):
# On macOS
brew install uv

# On Linux
curl -LsSf https://astral.sh/uv/install.sh | sh

# On Windows
powershell -ExecutionPolicy ByPass -c "irm https://astral.sh/uv/install.ps1 | iex"

Configuration

Create a .env file in the root directory with your API keys:

OPENAI_API_KEY=your_openai_api_key
ANTHROPIC_API_KEY=your_anthropic_api_key
GOOGLE_API_KEY=your_google_api_key
DEEPSEEK_API_KEY=your_deepseek_api_key

Usage

Using with Claude Desktop or Cursor

Add a server entry to your Claude Desktop configuration file or .cursor/mcp.json:

"mcpServers": {
  "llm-bridge": {
    "command": "uvx",
    "args": [
      "llm-bridge-mcp"
    ],
    "env": {
      "OPENAI_API_KEY": "your_openai_api_key",
      "ANTHROPIC_API_KEY": "your_anthropic_api_key",
      "GOOGLE_API_KEY": "your_google_api_key",
      "DEEPSEEK_API_KEY": "your_deepseek_api_key"
    }
  }
}

Troubleshooting

Common Issues

1. "spawn uvx ENOENT" Error

This error occurs when the system cannot find the uvx executable in your PATH. To resolve this:

Solution: Use the full path to uvx

Find the full path to your uvx executable:

# On macOS/Linux
which uvx

# On Windows
where.exe uvx

Then update your MCP server configuration to use the full path:

"mcpServers": {
  "llm-bridge": {
    "command": "/full/path/to/uvx",  // Replace with your actual path
    "args": [
      "llm-bridge-mcp"
    ],
    "env": {
      // ... your environment variables
    }
  }
}

License

This project is licensed under the MIT License - see the LICENSE file for details.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

llm_bridge_mcp-0.1.1.tar.gz (45.4 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

llm_bridge_mcp-0.1.1-py3-none-any.whl (4.6 kB view details)

Uploaded Python 3

File details

Details for the file llm_bridge_mcp-0.1.1.tar.gz.

File metadata

  • Download URL: llm_bridge_mcp-0.1.1.tar.gz
  • Upload date:
  • Size: 45.4 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: uv/0.6.6

File hashes

Hashes for llm_bridge_mcp-0.1.1.tar.gz
Algorithm Hash digest
SHA256 7d0deaca698fad9c49569ef29c212553ec19e79689713242581ed12fffbbde58
MD5 6df540bd0df571905b66d0baa695e43d
BLAKE2b-256 3f1e621c2b33d01e6a64b6e1e2da90825790674e47ca5865da2275c0d3e6c1ae

See more details on using hashes here.

File details

Details for the file llm_bridge_mcp-0.1.1-py3-none-any.whl.

File metadata

File hashes

Hashes for llm_bridge_mcp-0.1.1-py3-none-any.whl
Algorithm Hash digest
SHA256 0e4c946d7cadbdbbf227605bfd7de66f25ef96ac9a319b5cbf7d5a91efd703bb
MD5 b2b0e49ee61ab5f223ea2f8600fc8fc4
BLAKE2b-256 8493c98b0bd56a27753f276b740a32ae52020c53b8ee9b7c01f35355f0d83952

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page