Async OpenAI Responses API client with typed tool calling, lazy output parsing, and Azure/Databricks backends
Project description
llm-interaction
OpenAI Responses API client with typed tool calling, lazy output parsing, and Azure/Databricks/OpenRouter backends. Supports both async and sync (notebook-friendly) usage.
Install
pip install llm-interaction
# For Databricks backend:
pip install llm-interaction[databricks]
Quick Start
Loading Environment Variables
The library reads configuration from environment variables. Use python-dotenv to load them from a .env file:
from dotenv import load_dotenv
# Call this once at the start of your script/notebook
load_dotenv()
Example .env file:
LLM_INTERACTION_API_KEY=your-api-key
LLM_INTERACTION_ENDPOINT=https://your-resource.openai.azure.com
LLM_INTERACTION_MODEL=gpt-4o
If your environment variables are already set (e.g., in production), you can skip load_dotenv().
Sync (Notebook-Friendly)
from pathlib import Path
from llm_interaction import LLMInteraction
# Azure OpenAI (default)
llm = LLMInteraction(prompt_dir=Path("prompts"))
# Databricks
llm = LLMInteraction(prompt_dir=Path("prompts"), backend="databricks")
# OpenRouter
llm = LLMInteraction(
prompt_dir=Path("prompts"),
backend="openrouter",
api_key="your-openrouter-key",
model="openai/gpt-4",
)
# Use sync methods in notebooks (no await needed)
result = llm.sync_query(system="Be helpful", user="Hello")
print(result.text)
# Or with templates
result = llm.sync_query_template(
prompt_name="greeting",
variables={"name": "Alice"},
)
print(result.text)
# Or run an agentic loop
from llm_interaction import tool
@tool(stop=True)
def submit_answer(answer: str) -> str:
"""Submit the final answer."""
return "done"
result = llm.sync_agent_loop(
system="You are a helpful assistant.",
user="What is 2+2?",
tools=[submit_answer],
)
print(f"Tool calls: {result.tool_call_count}, Reason: {result.stop_reason}")
Async
from pathlib import Path
from llm_interaction import LLMInteraction
llm = LLMInteraction(prompt_dir=Path("prompts"))
# Use async methods in async contexts
result = await llm.query(system="Be helpful", user="Hello")
print(result.text)
# Or with templates
result = await llm.query_template(
prompt_name="greeting",
variables={"name": "Alice"},
)
print(result.text)
Output Parsing
Both query() and sync_query() return an LLMResponse. Parsing is lazy — call the method you need:
# Raw text
result.text
# JSON (extracts last ```json block, falls back to json_repair)
data = result.json()
# YAML
data = result.yaml()
# Scratchpad + JSON: splits reasoning text from structured data
scratchpad, data = result.scratchpad_json()
# scratchpad = "Let me analyze this step by step..."
# data = {"topics": ["ai", "ml"]}
# Scratchpad + YAML
scratchpad, data = result.scratchpad_yaml()
# Pydantic model (validates + auto-retries on failure)
from pydantic import BaseModel
class Analysis(BaseModel):
topics: list[str]
confidence: float
analysis = await result.parse(Analysis)
# On validation error, re-queries the LLM with the error message
# using previous_response_id for efficient context chaining
Tool Calling
@tool
def search(query: str, max_results: int = 10) -> list[dict]:
"""Search for documents.
Args:
query: The search query string
max_results: Maximum number of results to return
"""
return db.search(query, limit=max_results)
@tool(stop=True)
def submit(answer: str) -> str:
"""Submit the final answer."""
return "done"
result = await llm.agent_loop(
system="You are a research agent.",
user="Find info about quantum computing.",
tools=[search, submit],
)
Jinja Templates
Templates use the naming convention {name}_system.jinja and {name}_user.jinja:
# prompts/research_system.jinja
You are a {{ role }} assistant.
# prompts/research_user.jinja
Find information about {{ topic }}.
result = await llm.query_template(
prompt_name="research",
variables={"role": "research", "topic": "quantum computing"},
)
Context Injection
class WeatherAPI:
def get(self, city: str) -> dict:
return {"city": city, "temp_c": 22, "condition": "sunny"}
@tool
def get_weather(ctx: ToolContext[WeatherAPI], city: str) -> dict:
"""Get current weather for a city.
Args:
city: City name to look up
"""
return ctx.get(city)
weather_api = WeatherAPI()
result = await llm.agent_loop(
system="You are a helpful assistant with weather access.",
user="What's the weather in Oslo?",
tools=[get_weather],
context=[weather_api], # matched by type to ToolContext[WeatherAPI]
)
A single tool can use multiple contexts, each matched by type:
class WeatherAPI:
def get(self, city: str) -> dict:
return {"city": city, "temp_c": 22, "condition": "sunny"}
class UserPreferences:
def __init__(self, unit: str = "celsius"):
self.unit = unit
@tool
def get_weather(
weather: ToolContext[WeatherAPI],
prefs: ToolContext[UserPreferences],
city: str,
) -> str:
"""Get weather for a city in the user's preferred unit.
Args:
city: City name to look up
"""
data = weather.get(city)
if prefs.unit == "fahrenheit":
data["temp_f"] = data["temp_c"] * 9 / 5 + 32
return data
result = await llm.agent_loop(
system="You are a weather assistant.",
user="What's the weather in Oslo?",
tools=[get_weather],
context=[WeatherAPI(), UserPreferences(unit="fahrenheit")],
)
Environment Variables
Azure OpenAI (default backend):
LLM_INTERACTION_API_KEY=your-api-key
LLM_INTERACTION_ENDPOINT=https://your-resource.openai.azure.com
LLM_INTERACTION_MODEL=gpt-4o
Databricks (backend="databricks"):
LLM_INTERACTION_DATABRICKS_HOST=https://your-workspace.azuredatabricks.net
LLM_INTERACTION_MODEL=your-serving-endpoint
Databricks auth is handled automatically by WorkspaceClient:
- On-site (notebook): no setup needed
- Off-site (local dev): run
databricks auth login --host <your-host>first
OpenRouter (backend="openrouter"):
LLM_INTERACTION_API_KEY=your-openrouter-key
LLM_INTERACTION_MODEL=openai/gpt-4
License
MIT
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file llm_interaction-0.1.3.tar.gz.
File metadata
- Download URL: llm_interaction-0.1.3.tar.gz
- Upload date:
- Size: 23.8 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.12.11
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
5ab9ab075a05155d3e8f5f17baa828a1fc129cc66664f474ee97d99264c6e232
|
|
| MD5 |
34eacff90360c22f951a4b9abab582f4
|
|
| BLAKE2b-256 |
d3eb772aefabe10f4b866dbb3a340f53ddf33b6fdb5716e989073eab3dacffc9
|
File details
Details for the file llm_interaction-0.1.3-py3-none-any.whl.
File metadata
- Download URL: llm_interaction-0.1.3-py3-none-any.whl
- Upload date:
- Size: 16.0 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.12.11
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
9a6e4504aa7e7447d1d7c4394a60d9f3cc5ac27ae84c80a58676b5d9f8925b99
|
|
| MD5 |
8f2f95f52b0729cf2be4f56d45e0b283
|
|
| BLAKE2b-256 |
23bde6d798dc084776970022e67e71401f0577620d83303adf2bd8c308d9e996
|