Skip to main content

Universal interface for different LLM providers

Project description

univllm

PyPI version

A universal Python package that provides a standardised interface for different LLM providers including OpenAI, Anthropic, Deepseek, Mistral, and Gemini.

Features

  • Universal Interface: Single API to interact with multiple LLM providers
  • Auto-Detection: Automatically detect the appropriate provider based on model name
  • Streaming Support: Stream completions from all supported providers
  • MCP Tool Calling: Compatible with Model Context Protocol (MCP) for function/tool calling
  • Model Capabilities: Query model capabilities like context window, function calling support, etc.
  • Error Handling: Comprehensive error handling with provider-specific exceptions
  • Async Support: Fully asynchronous API for better performance

Supported Providers

  • OpenAI: GPT-4o, GPT-5.x & GPT-5.2 family models
  • Anthropic: Claude 3.x, 4.x & 4.5 family models
  • Deepseek: DeepSeek V3.2, Chat, Reasoner, Coder & VL models
  • Mistral: Mistral Large 3, Ministral 3, Magistral, Codestral & specialized models
  • Gemini: Google Gemini 1.5, 2.0 & 2.5 family models

Supported Model Prefixes

Provider auto-detection uses simple prefix matching (see SUPPORTED_MODELS lists). Known prefixes are used as hints for provider selection and capability metadata, but request models accept arbitrary model strings so new provider releases are not rejected before an API call. Provider-specific suffixes or date/version tags (e.g. -20240229, -latest, -0125, minor patch tags) are allowed.

Provider Accepted Prefixes (Exact / Prefix Match) Notes
OpenAI gpt-5.2, gpt-5.1, gpt-5, gpt-5-mini, gpt-5-nano, gpt-5-codex, gpt-oss-120b, gpt-oss-20b, gpt-vision-1, gpt-4o, gpt-4 GPT-5.2 is the latest flagship model (Dec 2025). Any extended suffix (e.g. gpt-5.2-2025-12-11) will pass if it starts with a listed prefix.
Anthropic claude-opus-4-5, claude-sonnet-4-5, claude-haiku-4-5, claude-opus-4.1, claude-sonnet-4-, claude-haiku-4-, claude-opus-4-, claude-code, claude-3-7-sonnet-, claude-3-5-sonnet- Claude 4.5 series launched Sep-Nov 2025. Older variants (e.g. dated claude-3-* forms) can be added by extending the list in supported_models.py.
Deepseek deepseek-chat, deepseek-reasoner, deepseek-coder, deepseek-vl, deepseek-v3 DeepSeek V3.2 models. deepseek-reasoner for advanced reasoning tasks. deepseek-vl for vision-language.
Mistral mistral-large-3, mistral-medium-3, mistral-small-3, ministral-3-, magistral-medium-, magistral-small-, codestral-, devstral-, voxtral-, mistral-ocr-, ocr-3- Mistral Large 3 (Dec 2025) flagship multimodal model. Ministral for edge, Codestral for code generation.
Gemini gemini-2.5-pro, gemini-2.5-flash, gemini-2.0-flash, gemini-1.5-pro, gemini-1.5-flash Google's Gemini models across multiple versions.

Note: If a model name is not yet in the prefix lists, you can still pass it directly when you also know the provider. Adding new prefixes mainly improves auto-detection and capability lookup.

Installation

pip install univllm

Quick Start

import asyncio
from univllm import UniversalLLMClient


async def main():
    client = UniversalLLMClient()

    # Auto-detects provider based on model name
    response = await client.complete(
        messages=[{"role": "user", "content": "What is the capital of France?"}],
        model="gpt-5.2"
    )

    print(response.content)


asyncio.run(main())

Configuration

Set your API keys as environment variables:

export OPENAI_API_KEY="your-openai-key"
export ANTHROPIC_API_KEY="your-anthropic-key"
export DEEPSEEK_API_KEY="your-deepseek-key"
export MISTRAL_API_KEY="your-mistral-key"
export GEMINI_API_KEY="your-gemini-key"

Or pass them directly:

from univllm import UniversalLLMClient, ProviderType

client = UniversalLLMClient(
    provider=ProviderType.OPENAI,
    api_key="your-api-key"
)

Usage Examples

Basic Completion

import asyncio
from univllm import UniversalLLMClient


async def main():
    client = UniversalLLMClient()

    response = await client.complete(
        messages=[
            {"role": "system", "content": "You are a helpful assistant."},
            {"role": "user", "content": "Explain quantum computing briefly."}
        ],
        model="gpt-5.2",
        max_tokens=150,
        temperature=0.7
    )

    print(f"Response: {response.content}")
    print(f"Provider: {response.provider}")
    print(f"Model: {response.model}")
    print(f"Usage: {response.usage}")


asyncio.run(main())

Streaming Completion

import asyncio
from univllm import UniversalLLMClient


async def main():
    client = UniversalLLMClient()

    async for chunk in client.stream_complete(
            messages=[{"role": "user", "content": "Tell me a short story about a robot."}],
            model="claude-sonnet-4-5",
            max_tokens=200
    ):
        print(chunk, end="", flush=True)


asyncio.run(main())

Model Capabilities

import asyncio
from univllm import UniversalLLMClient


async def main():
    client = UniversalLLMClient()

    # Get capabilities for a specific model
    capabilities = client.get_model_capabilities("gpt-5.2")

    print(f"Supports function calling: {capabilities.supports_function_calling}")
    print(f"Context window: {capabilities.context_window}")
    print(f"Max tokens: {capabilities.max_tokens}")

    # Get all supported models
    all_models = client.get_supported_models()
    for provider, models in all_models.items():
        print(f"{provider}: {len(models)} models")


asyncio.run(main())

Tool Calling (MCP Compatible)

univllm supports function/tool calling following the Model Context Protocol (MCP) format:

import asyncio
from univllm import UniversalLLMClient, ToolDefinition


async def main():
    client = UniversalLLMClient()

    # Define a tool using MCP format
    weather_tool = ToolDefinition(
        name="get_weather",
        description="Get current weather for a location",
        input_schema={
            "type": "object",
            "properties": {
                "location": {
                    "type": "string",
                    "description": "City name or zip code"
                }
            },
            "required": ["location"]
        }
    )

    # Request with tools
    response = await client.complete(
        messages=[{"role": "user", "content": "What's the weather in Paris?"}],
        model="gpt-4o",
        tools=[weather_tool],
        tool_choice="auto"  # Let the model decide when to use tools
    )

    # Check if model wants to call a tool
    if response.tool_calls:
        tool_call = response.tool_calls[0]
        print(f"Tool: {tool_call.name}")
        print(f"Arguments: {tool_call.arguments}")
        
        # Execute the tool and get result
        # ... your tool execution logic ...
        
        # Continue conversation with tool result
        # ... send tool result back to model ...


asyncio.run(main())

You can also pass tools as dictionaries:

tools = [
    {
        "name": "calculate",
        "description": "Perform arithmetic calculations",
        "input_schema": {
            "type": "object",
            "properties": {
                "expression": {"type": "string"}
            },
            "required": ["expression"]
        }
    }
]

response = await client.complete(
    messages=[{"role": "user", "content": "Calculate 15 * 23"}],
    model="gpt-4o",
    tools=tools
)

See examples_tool_calling.py for more comprehensive examples.

Multiple Providers

import asyncio
from univllm import UniversalLLMClient
from univllm.models import ProviderType


async def main():
    client = UniversalLLMClient()

    question = "What is machine learning?"

    # Use separate clients or call set_provider() when switching providers.
    openai_client = UniversalLLMClient()
    anthropic_client = UniversalLLMClient()

    openai_response = await openai_client.complete(
        messages=[{"role": "user", "content": question}],
        model="gpt-5.2"
    )

    anthropic_response = await anthropic_client.complete(
        messages=[{"role": "user", "content": question}],
        model="claude-sonnet-4-5"
    )

    print(f"OpenAI: {openai_response.content[:100]}...")
    print(f"Anthropic: {anthropic_response.content[:100]}...")


asyncio.run(main())

API Reference

UniversalLLMClient

Main client class for interacting with LLM providers.

Methods

  • complete(): Generate a completion
  • stream_complete(): Generate a streaming completion
  • get_model_capabilities(): Get model capabilities
  • get_supported_models(): Get supported models for all providers
  • set_provider(): Set or change the provider

Models

  • CompletionRequest: Request object for completions
  • CompletionResponse: Response object from completions
  • ModelCapabilities: Information about model capabilities
  • Message: Individual message in a conversation

Providers

  • ProviderType: Enum of supported providers
  • BaseLLMProvider: Base class for provider implementations

Exceptions

  • UniversalLLMError: Base exception
  • ProviderError: Provider-related errors
  • ModelNotSupportedError: Unsupported model errors
  • AuthenticationError: Authentication failures
  • ConfigurationError: Configuration issues

Development

git clone https://github.com/nihilok/univllm.git
cd univllm
pip install -e ".[dev]"

Run tests:

pytest

Licence

MIT Licence

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distributions

No source distribution files available for this release.See tutorial on generating distribution archives.

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

univllm-0.2.0-py3-none-any.whl (24.8 kB view details)

Uploaded Python 3

File details

Details for the file univllm-0.2.0-py3-none-any.whl.

File metadata

  • Download URL: univllm-0.2.0-py3-none-any.whl
  • Upload date:
  • Size: 24.8 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.10.18

File hashes

Hashes for univllm-0.2.0-py3-none-any.whl
Algorithm Hash digest
SHA256 5534a58ecccf541120fcd47f97f9a7d88ec870015f9a7d57e9cff9ee14ed989b
MD5 020b4709a136f14d9f778f27f5c4cfd8
BLAKE2b-256 7748ccb4f526aceaa89722f83af53e3656e496b13348eec47fbc66e99c6e859e

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page