Skip to main content

This is an unofficial wrapper of OpenRouter.

Project description

Introduction

Welcome to openrouter-provider, an unofficial Python wrapper for the OpenRouter API. This library lets you easily integrate with OpenRouter models, manage chat sessions, process images, and call tools within your Python application.

Features

  • Simple chat interface with system, user, assistant, and tool roles
  • Automatic image resizing and Base64 encoding
  • Built-in tool decorator for defining custom functions

Installation

From PyPI

pip3 install openrouter-provider

From Source

git clone https://github.com/yourusername/openrouter-provider.git
cd openrouter-provider
pip3 install .

Configuration

  1. Create a .env file in your project root.

  2. Add your OpenRouter API key:

    OPENROUTER_API_KEY=your_api_key_here
    

Usage

Basic chat bot

Chat history is automatically sent, by Chatbot_manager. If you want to delete chat history, use clear_memory method.

from OpenRouterProvider.Chatbot_manager import Chat_message, Chatbot_manager
from OpenRouterProvider.LLMs import gpt_4o_mini

# Declare chat bot
ai = Chatbot_manager(system_prompt="Please answer in English.")

# Send query
query = Chat_message(text="Introduce yourself, please.")
response = ai.invoke(model=gpt_4o_mini, query=query)
print(response.text)

# Send next query. Chatbot_manager automatically handle chat history.
query = Chat_message(text="Tell me a short story.")
response = ai.invoke(model=gpt_4o_mini, query=query)
print(response.text)

# Print all chat history
ai.print_memory()  

# Delete all chat history
ai.clear_memory()

Chat bot with images

You can use images in the chat.

from OpenRouterProvider.Chatbot_manager import Chat_message, Chatbot_manager
from OpenRouterProvider.LLMs import gpt_4o_mini
from PIL import Image

dog = Image.open("dog.jpg")
cat = Image.open("cat.jpg")

# Send query with images
ai = Chatbot_manager(system_prompt="Please answer in English.")
query = Chat_message(text="What can you see in the images?", images=[dog, cat])
response = ai.invoke(model=gpt_4o_mini, query=query)
print(response.text) 

With tools

Use the @tool_model decorator to expose Python functions as callable tools in the chat. Tools are automatically processed by Chat_manager, so you don't need to care it.

from OpenRouterProvider.Chatbot_manager import Chat_message, Chatbot_manager
from OpenRouterProvider.LLMs import gpt_4o_mini
from OpenRouterProvider.Tool import tool_model

@tool_model
def get_user_info():
    """
    Return user's personal info: name, age, and address.
    """
    return "name: Alice\nage: 30\naddress: Wonderland"

ai = Chatbot_manager(system_prompt="Please answer in English.", tools=[get_user_info])
query = Chat_message(text="What is the name, age, address of the user?")
response = ai.invoke(model=gpt_4o_mini, query=query)
ai.print_memory()

Advanced Usage

Prebuilt and Custom Model Usage

You can use prebuilt models defined or declare your own custom models easily. This library provides many ready-to-use models from OpenAI, Anthropic, Google, and others.

from OpenRouterProvider.Chatbot_manager import Chat_message, Chatbot_manager
from OpenRouterProvider.LLMs import gpt_4o, claude_3_7_sonnet

# Use OpenAI GPT-4o
ai = Chatbot_manager(system_prompt="Please answer in English.")
query = Chat_message(text="Tell me a joke.")
response = ai.invoke(model=gpt_4o, query=query)
print(response.text)

# Use Anthropic Claude 3.7 Sonnet
query = Chat_message(text="Summarize the story of Hamlet.")
response = ai.invoke(model=claude_3_7_sonnet, query=query)
print(response.text)

Available prebuilt models include:

OpenAI

  • gpt_4o
  • gpt_4o_mini
  • gpt_4_1
  • gpt_4_1_mini
  • gpt_4_1_nano
  • o4_mini
  • o4_mini_high
  • o3

Anthropic

  • claude_3_7_sonnet
  • claude_3_7_sonnet_thinking
  • claude_3_5_haiku

Google

  • gemini_2_0_flash
  • gemini_2_0_flash_free
  • gemini_2_5_flash
  • gemini_2_5_flash_thinking
  • gemini_2_5_pro

Deepseek

  • deepseek_v3_free
  • deepseek_v3
  • deepseek_r1_free
  • deepseek_r1

xAI

  • grok_3_mini
  • grok_3

Microsoft

  • mai_ds_r1_free

Others

  • llama_4_maverick_free
  • llama_4_scout
  • mistral_small_3_1_24B_free

All of them are instances of LLMModel, which includes cost and model name settings.

Using Custom Models

You can define and use your own custom model if it's available on OpenRouter.

from OpenRouterProvider.Chatbot_manager import Chat_message, Chatbot_manager
from OpenRouterProvider.LLMs import LLMModel

# Declare a custom model
my_model = LLMModel(
    name="my-org/my-custom-model",  # Model name for OpenRouter
    input_cost=0.5,                 # Optional: cost per 1M input tokens
    output_cost=2.0                 # Optional: cost per 1M output tokens
)

# Use the custom model
ai = Chatbot_manager(system_prompt="Please answer in English.")
query = Chat_message(text="Explain black holes simply.")
response = ai.invoke(model=my_model, query=query)
print(response.text)

You only need to know the model name as used on OpenRouter. input_cost and output_cost are optional and currently, they are not used in this library. Please wait the future update.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

openrouter_provider-0.0.8.tar.gz (11.1 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

openrouter_provider-0.0.8-py3-none-any.whl (10.7 kB view details)

Uploaded Python 3

File details

Details for the file openrouter_provider-0.0.8.tar.gz.

File metadata

  • Download URL: openrouter_provider-0.0.8.tar.gz
  • Upload date:
  • Size: 11.1 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.12.2

File hashes

Hashes for openrouter_provider-0.0.8.tar.gz
Algorithm Hash digest
SHA256 098252fb724232ed4c0262446e07dc1a912815e2ab3be156de5be14593974655
MD5 51d1cc1d6656898803b39b8c0b982608
BLAKE2b-256 08685f0637a4f5ad5dd37a7e2175e28068dc01545153f23937bbc247fdbae805

See more details on using hashes here.

File details

Details for the file openrouter_provider-0.0.8-py3-none-any.whl.

File metadata

File hashes

Hashes for openrouter_provider-0.0.8-py3-none-any.whl
Algorithm Hash digest
SHA256 63edf600d229b5f8a1fef30ea3b7b5e49d3d82316869a9ae8f6318241d2ed6dc
MD5 3a8b9a6924b42bc89ccb45719e1525bb
BLAKE2b-256 eaba6ed0c49d60c03573cfa28d3add0e571a6f5aaefe8de4eb0a5d6d4e65fc51

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page