Minimal LLM client getter for OpenAI Responses + OpenAI-compatible Chat Completions.
Project description
kantan-llm 😺✨
A tiny Python library that removes the boring boilerplate (keys/URLs/provider selection) so you can call LLMs with a single get_llm() 💨
Big idea: set env vars for the providers/models you use, then just do get_llm("model-name") and it “just connects” 😺✨
Supported providers (roughly) 🌍
- OpenAI (Responses)
- Anthropic (Claude via OpenAI-compatible SDK)
- OpenRouter (OpenAI-compatible Chat)
- Google (Gemini via OpenAI-compatible Chat)
- LMStudio / Ollama / any OpenAI-compatible Chat
Install 📦
pip install kantan-llm
Quickstart 🚀
OpenAI (Responses API is the source of truth)
export OPENAI_API_KEY="sk-..."
from kantan_llm import get_llm
llm = get_llm("gpt-4.1-mini")
res = llm.responses.create(input="Say hi in one short line.")
print(res.output_text)
llm is OpenAI SDK compatible (unknown attributes delegate to the underlying client).
OpenAI-compatible (Chat Completions is the source of truth)
LMStudio (example: openai/gpt-oss-20b)
export LMSTUDIO_BASE_URL="http://192.168.11.16:1234" # `/v1` is optional
from kantan_llm import get_llm
llm = get_llm("openai/gpt-oss-20b", provider="lmstudio")
cc = llm.chat.completions.create(messages=[{"role": "user", "content": "Return exactly: OK"}], max_tokens=16)
print(cc.choices[0].message.content)
Ollama (example)
export OLLAMA_BASE_URL="http://localhost:11434" # `/v1` is optional
from kantan_llm import get_llm
llm = get_llm("llama3.2", provider="ollama")
cc = llm.chat.completions.create(messages=[{"role": "user", "content": "Return exactly: OK"}], max_tokens=16)
print(cc.choices[0].message.content)
Anthropic (Claude via OpenAI-compatible SDK)
export CLAUDE_API_KEY="sk-ant-..."
from kantan_llm import get_llm
llm = get_llm("claude-3-5-sonnet-latest") # if `CLAUDE_API_KEY` exists -> provider=anthropic (inferred)
cc = llm.chat.completions.create(messages=[{"role": "user", "content": "Return exactly: OK"}], max_tokens=16)
print(cc.choices[0].message.content)
OpenRouter (includes Claude, etc.)
export OPENROUTER_API_KEY="..."
from kantan_llm import get_llm
llm = get_llm("anthropic/claude-3.5-sonnet", provider="openrouter") # explicit is recommended (Anthropic takes precedence)
cc = llm.chat.completions.create(messages=[{"role": "user", "content": "Return exactly: OK"}], max_tokens=16)
print(cc.choices[0].message.content)
Google (Gemini via an OpenAI-compatible endpoint)
export GOOGLE_API_KEY="..."
from kantan_llm import get_llm
llm = get_llm("gemini-2.0-flash")
cc = llm.chat.completions.create(messages=[{"role": "user", "content": "Return exactly: OK"}], max_tokens=16)
print(cc.choices[0].message.content)
Provider rules 🧭
gpt-*→openaigemini-*→googleclaude-*→anthropic(ifCLAUDE_API_KEYis set) →openrouter(ifOPENROUTER_API_KEYis set) → otherwisecompat- If the model name is not recognizable, it picks the first available provider by env vars:
lmstudio→ollama→openrouter→anthropic→google
Explicit provider 🎯
from kantan_llm import get_llm
llm = get_llm("gpt-4.1-mini", provider="openai")
Fallback (order = priority) 🧯
from kantan_llm import get_llm
llm = get_llm("gpt-4.1-mini", providers=["openai", "lmstudio", "openrouter"])
Tracing / Tracer 🧵
By default, get_llm() enables a simple tracer that prints input/output (colorized) for each LLM call.
from kantan_llm import get_llm
from kantan_llm.tracing import trace
llm = get_llm("gpt-4.1-mini")
with trace("workflow"):
llm.responses.create(input="Say hi.")
More: docs/tracing.md
Search (SQLite) 🔎
Use SQLiteTracer as a lightweight search backend for traces/spans.
from kantan_llm.tracing import SpanQuery, TraceQuery
from kantan_llm.tracing.processors import SQLiteTracer
tracer = SQLiteTracer("traces.sqlite3")
traces = tracer.search_traces(query=TraceQuery(keywords=["hello"], limit=10))
spans = tracer.search_spans(query=SpanQuery(keywords=["hello"], limit=10))
More: docs/search.md
Tutorial: docs/tutorial_trace_analysis.md
Examples 📚
examples/tracing_basic.pyexamples/search_sqlite.py
Environment variables 🔐
- OpenAI
OPENAI_API_KEY(required)OPENAI_BASE_URL(optional)
- Generic compatible (
compat)KANTAN_LLM_BASE_URL(required)KANTAN_LLM_API_KEY(optional; falls back to a dummy value)
- LMStudio
LMSTUDIO_BASE_URL(required)
- Ollama
OLLAMA_BASE_URL(required)
- OpenRouter
OPENROUTER_API_KEY(required)
- Anthropic
CLAUDE_API_KEY(required)CLAUDE_BASE_URL(optional)
- Google
GOOGLE_API_KEY(required)GOOGLE_BASE_URL(optional)
Error example 💥
- Missing OpenAI key:
python -c 'from kantan_llm import get_llm; get_llm(\"gpt-4.1-mini\")'→[kantan-llm][E2] Missing OPENAI_API_KEY for provider: openai
Tests 🧪
Live integration tests (real APIs) are opt-in:
KANTAN_LLM_RUN_LIVE_TESTS=1 pytest -q -m integration
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file kantan_llm-0.1.6.tar.gz.
File metadata
- Download URL: kantan_llm-0.1.6.tar.gz
- Upload date:
- Size: 29.6 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.12.10
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
d95b39e4ee4fcdbdb8fdc183b5c80c0fe713d6c52470fc0f0c38a557dc17d242
|
|
| MD5 |
81270a1b3b6309972c2581887abd8cc0
|
|
| BLAKE2b-256 |
6ee4d033905507efedf3341f21f4611bab03b1737bbcf51d24ea8315ee24b010
|
File details
Details for the file kantan_llm-0.1.6-py3-none-any.whl.
File metadata
- Download URL: kantan_llm-0.1.6-py3-none-any.whl
- Upload date:
- Size: 29.6 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.12.10
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
1a12c82790bd242f9ec26b28fd3416dd43ded512ca2a776cd34e49021e641b60
|
|
| MD5 |
124afb14f1ce79b271bdf2d0e8a84200
|
|
| BLAKE2b-256 |
1302f200e995018ed302f558fa504f05715171b81072bdfebc7779ca4c1ee405
|