Skip to main content

MBX AI SDK

Project description

MBX AI

A Python library for building AI applications with LLMs.

Features

  • OpenRouter Integration: Connect to various LLM providers through OpenRouter
  • Tool Integration: Easily integrate tools with LLMs using the Model Context Protocol (MCP)
  • Structured Output: Get structured, typed responses from LLMs
  • Chat Interface: Simple chat interface for interacting with LLMs
  • FastAPI Server: Built-in FastAPI server for tool integration

Installation

pip install mbxai

Quick Start

Basic Usage

from mbxai import OpenRouterClient

# Initialize the client
client = OpenRouterClient(api_key="your-api-key")

# Chat with an LLM
response = await client.chat([
    {"role": "user", "content": "Hello, how are you?"}
])
print(response.choices[0].message.content)

Using Tools

from mbxai import OpenRouterClient, ToolClient
from pydantic import BaseModel

# Define your tool's input and output models
class CalculatorInput(BaseModel):
    a: float
    b: float

class CalculatorOutput(BaseModel):
    result: float

# Create a calculator tool
async def calculator(input: CalculatorInput) -> CalculatorOutput:
    return CalculatorOutput(result=input.a + input.b)

# Initialize the client with tools
client = ToolClient(OpenRouterClient(api_key="your-api-key"))
client.add_tool(calculator)

# Use the tool in a chat
response = await client.chat([
    {"role": "user", "content": "What is 2 + 3?"}
])
print(response.choices[0].message.content)

Using MCP (Model Context Protocol)

from mbxai import OpenRouterClient, MCPClient
from mbxai.mcp import MCPServer
from mcp.server.fastmcp import FastMCP
from pydantic import BaseModel

# Define your tool's input and output models
class CalculatorInput(BaseModel):
    a: float
    b: float

class CalculatorOutput(BaseModel):
    result: float

# Create a FastMCP instance
mcp = FastMCP("calculator-service")

# Create a calculator tool
@mcp.tool()
async def calculator(argument: CalculatorInput) -> CalculatorOutput:
    return CalculatorOutput(result=argument.a + argument.b)

# Start the MCP server
server = MCPServer("calculator-service")
await server.add_tool(calculator)
await server.start()

# Initialize the MCP client
client = MCPClient(OpenRouterClient(api_key="your-api-key"))
await client.register_mcp_server("calculator-service", "http://localhost:8000")

# Use the tool in a chat
response = await client.chat([
    {"role": "user", "content": "What is 2 + 3?"}
])
print(response.choices[0].message.content)

Development

Setup

  1. Clone the repository:
git clone https://github.com/yourusername/mbxai.git
cd mbxai
  1. Create a virtual environment:
python -m venv .venv
source .venv/bin/activate  # On Windows: .venv\Scripts\activate
  1. Install dependencies:
pip install -e ".[dev]"

Running Tests

pytest tests/

License

MIT License

Project details


Release history Release notifications | RSS feed

This version

1.2.0

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

mbxai-1.2.0.tar.gz (22.3 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

mbxai-1.2.0-py3-none-any.whl (33.4 kB view details)

Uploaded Python 3

File details

Details for the file mbxai-1.2.0.tar.gz.

File metadata

  • Download URL: mbxai-1.2.0.tar.gz
  • Upload date:
  • Size: 22.3 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.12.9

File hashes

Hashes for mbxai-1.2.0.tar.gz
Algorithm Hash digest
SHA256 a8807c21b57e0b7d6c7e8cbfc9d8807b36b990d17c9fbe81be62d9c242559354
MD5 2d0fe1b4914a2ccaaf00763e4b8eeef8
BLAKE2b-256 d9ff8b1894ace83ae477bb7e53811dc0df71fddf307f7144d8fa056033cf3cac

See more details on using hashes here.

File details

Details for the file mbxai-1.2.0-py3-none-any.whl.

File metadata

  • Download URL: mbxai-1.2.0-py3-none-any.whl
  • Upload date:
  • Size: 33.4 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.12.9

File hashes

Hashes for mbxai-1.2.0-py3-none-any.whl
Algorithm Hash digest
SHA256 fb51b3e8a6b13b38bf57a5d6e7513c5c3e97f063c5a8d95e705a06242aee3c3b
MD5 26813f1e85514c7e035b99759f644f50
BLAKE2b-256 db799224eb7af471a7839e5371794a044a8281904adc6748e0ad04bf5cc00d2b

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page