Skip to main content

CLI to manage the OpenAI Responses Server that bridges chat completions to responses API calls

This project has been archived.

The maintainers of this project have marked this project as archived. No new releases are expected.

Project description

🚀 openai-responses-server

A plug-and-play server that speaks OpenAI’s Responses API—no matter which AI backend you’re running.

Ollama? vLLM? LiteLLM? Even OpenAI itself?
This server bridges them all to the OpenAI ChatCompletions & Responses API interface.

In plain words:
👉 Want to run OpenAI’s Coding Assistant (Codex) or other OpenAI API clients against your own models?
👉 Want to experiment with self-hosted LLMs but keep OpenAI’s API compatibility?

This project makes it happen.
It handles stateful chat, tool calls, and future features like file search & code interpreter—all behind a familiar OpenAI API.

✨ Why use this?

✅ Acts as a drop-in replacement for OpenAI’s Responses API.
✅ Lets you run any backend AI (Ollama, vLLM, Groq, etc.) with OpenAI-compatible clients.
✅ Supports OpenAI’s new Coding Assistant / Codex that requires Responses API.
✅ Built for innovators, researchers, OSS enthusiasts.
✅ Enterprise-ready: scalable, reliable, and secure for production workloads.

🏗️ Quick Install

Latest release on PyPI:

pip install openai-responses-server

Or install from source:

uv venv
uv pip install .
uv pip install -e ".[dev]"  # dev dependencies

Run the server:

uv run src/openai_responses_server/cli.py start

Works great with docker-compose.yaml for Codex + your own model.

🔥 What’s in & what’s next?

✅ Done 📝 Coming soon

  • ✅ Tool call support .env file support
  • ✅ Manual & pipeline tests
  • ✅ Docker image build
  • ✅ PyPI release
  • 📝 Persistent state (not just in-memory)
  • ✅ CLI validation
  • 📝 hosted tools:
    • 📝 MCPs support
    • 📝 Web search: crawl4ai
    • 📝 File upload + search: graphiti
    • 📝 Code interpreter
    • 📝 Computer use APIs

🛠️ Configure

Minimal config to connect your AI backend:

OPENAI_BASE_URL_INTERNAL=http://localhost:11434  # Ollama, vLLM, Groq, etc.
OPENAI_BASE_URL=http://localhost:8080            # This server’s endpoint
OPENAI_API_KEY=sk-mockapikey123456789            # Mock key tunneled to backend

Server binding:

API_ADAPTER_HOST=0.0.0.0
API_ADAPTER_PORT=8080

Optional logging:

LOG_LEVEL=INFO
LOG_FILE_PATH=./log/api_adapter.log

💬 I’d love your support!

If you think this is cool:
⭐ Star the repo.
🐛 Open an issue if something’s broken.
🤝 Suggest a feature or submit a pull request!

This is early-stage but already usable in real-world demos.
Let’s build something powerful—together.

📚 Citations & inspirations

Referenced projects

Cite this project

Code citation

@software{openai-responses-server,
  author = {TeaBranch},
  title = {openai-responses-server: Open-source server bridging any AI provider to OpenAI’s Responses API},
  year = {2025},
  publisher = {GitHub},
  journal = {GitHub Repository},
  howpublished = {\url{https://github.com/teabranch/openai-responses-server}},
  commit = {use the commit hash you’re working with}
}

Text citation

TeaBranch. (2025). openai-responses-server: Open-source server the serves any AI provider with OpenAI ChatCompletions as OpenAI's Responses API and hosted tools. [Computer software]. GitHub. https://github.com/teabranch/openai-responses-server

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

openai_responses_server-0.2.4.tar.gz (18.9 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

openai_responses_server-0.2.4-py3-none-any.whl (14.5 kB view details)

Uploaded Python 3

File details

Details for the file openai_responses_server-0.2.4.tar.gz.

File metadata

  • Download URL: openai_responses_server-0.2.4.tar.gz
  • Upload date:
  • Size: 18.9 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.12.9

File hashes

Hashes for openai_responses_server-0.2.4.tar.gz
Algorithm Hash digest
SHA256 4e1aa6787259bf8ea2702b09dee4d9b78cab441ef36e2eb6bf4172dd5ddc1046
MD5 78679ea4a5f28ceebbec206cb6d69f2d
BLAKE2b-256 a6d105adcc9d04d7bf74d4f4c2094b72130238df05af0740124cffea25c41228

See more details on using hashes here.

Provenance

The following attestation bundles were made for openai_responses_server-0.2.4.tar.gz:

Publisher: publish.yml on teabranch/openai-responses-server

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file openai_responses_server-0.2.4-py3-none-any.whl.

File metadata

File hashes

Hashes for openai_responses_server-0.2.4-py3-none-any.whl
Algorithm Hash digest
SHA256 5828b7f5191d25421a1588ce3aa4cadd0fee495e37ad75855a9767bee709a64c
MD5 041f76d815f77f8c7ac7882e64d6ffb6
BLAKE2b-256 1a8a9895e9382563d3e095e7bc2f848739b8a811c7833c49c238df8f13e59fe4

See more details on using hashes here.

Provenance

The following attestation bundles were made for openai_responses_server-0.2.4-py3-none-any.whl:

Publisher: publish.yml on teabranch/openai-responses-server

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page