Skip to main content

Moon for Claude: run Claude Code on external LLMs via LiteLLM

Project description

MoonClaude 🌙

Moon for Claude: Run Claude Code fully autonomously on external LLMs (e.g. Qwen, Gemini, MiniMax, OpenAI) through a LiteLLM proxy, while enjoying zero-downtime hot reloading and an Antigravity-grade persistent AI memory engine.


🚀 Features

  • Free OpenRouter Models: Interactively fetch and route your prompts through any of the 25+ completely free models on OpenRouter, or configure your own premium API keys (Gemini, Groq, OpenAI).
  • True Zero-Downtime Hot Reloading: Press Alt+M directly in the proxy terminal to instantaneously swap the active model. MoonClaude dynamically intercepts the routing under the hood—no proxy restarts, no lost context in the CLI, no downtime.
  • Antigravity-Grade Project Memory:
    • Auto-discovers and enforces MOONCLAUDE.md global project instructions into Claude Code's system prompts.
    • Generates cross-session, highly structured summarization context representing your recent AI coding sessions, intelligently managing your AI's working context without maxing out open files.
    • Easily browse transcripts natively with moon history.

📦 Installation & Setup

1. Prerequisites

If you haven't already, you need Node.js installed to run Anthropic's official Claude Code CLI.

# Install Claude Code globally via npm
npm install -g @anthropic-ai/claude-code

2. Install MoonClaude

MoonClaude is available directly on PyPI!

pip install moonclaude

3. Initialize Settings

Run the interactive setup wizard to map your API keys (OpenRouter, Gemini, Groq) and select your preferred chat models.

moon setup

🏃 Quick Start

The easiest way to launch the proxy and jump straight into Claude Code is a single command. Open your project folder and run:

moon launch

(If you haven't run setup yet, it will guide you through it first!)

Manual Control: If you prefer running the Proxy and Claude independently, open two terminals:

  • Terminal 1: Run moon start (Starts the LiteLLM routing proxy)
  • Terminal 2: Run moon chat (Launches Claude Code seamlessly attached to the proxy)

🎛️ Core Commands

moon setup                  Interactive wizard for API keys and preferred models
moon launch                 Quick launch (starts proxy in background, then launches Claude)
moon chat                   Launch Claude connected to the proxy
moon start                  Start LiteLLM proxy in the foreground with live token monitoring
moon switch                 Fast CLI model switcher 
moon history                Interactive rich browser for your past conversation logs
moon memory                 Detailed control over the project memory context engine
moon status                 Display system health, active models, and paths
moon                        Show CLI help and commands

Compatibility: MoonClaude preserves the old claude-ext ... legacy command aliases. Your legacy ~/.claude-ext config will automatically migrate to ~/.moonclaude/ at launch!


♻️ Zero-Downtime Hot Reloading

When you are deep in a Claude Code debug session, discovering your chosen model is failing can be frustrating.

Instead of dropping context:

  1. Keep the proxy running in the foreground (moon start).
  2. Press Alt+M.
  3. Select a new LLM from your interactive UI overlay.
  4. MoonClaude dynamically pre-warms its internal router injection without dropping the proxy connection. Your next prompt in the Claude Code terminal instantly bridges over to the new model!

🧠 Memory Engine

MoonClaude injects long-term memory into Claude Code.

  1. MOONCLAUDE.md: Place this file in your project root. MoonClaude automatically reads it and appends it to your AI's persistent context across all sessions.
  2. Context Compression: When you boot moon chat, MoonClaude evaluates your last 20 sessions and constructs an elegant, compressed timeline of files you've modified, goals you've accomplished, and context you shouldn't forget.
  3. Session Replay: Type moon history to explore Rich-formatted logs of all of your previous AI code iterations.

License

MIT

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

moonclaude-2.0.0.tar.gz (37.2 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

moonclaude-2.0.0-py3-none-any.whl (39.6 kB view details)

Uploaded Python 3

File details

Details for the file moonclaude-2.0.0.tar.gz.

File metadata

  • Download URL: moonclaude-2.0.0.tar.gz
  • Upload date:
  • Size: 37.2 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.13.5

File hashes

Hashes for moonclaude-2.0.0.tar.gz
Algorithm Hash digest
SHA256 450efafd3cb2879c4173a2c038516f8ca185dd6387913b9d429eef310e057bed
MD5 e223c96f18284168299afce3b14be4f1
BLAKE2b-256 a06d22bdc59e9349797807bf81a81edd96f983a828353de1277ea7c54645faa7

See more details on using hashes here.

File details

Details for the file moonclaude-2.0.0-py3-none-any.whl.

File metadata

  • Download URL: moonclaude-2.0.0-py3-none-any.whl
  • Upload date:
  • Size: 39.6 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.13.5

File hashes

Hashes for moonclaude-2.0.0-py3-none-any.whl
Algorithm Hash digest
SHA256 c59b18d802b402310a8f29d0bf33270b4e16598e410378bbfab877a70800c263
MD5 b8bf005e8d63d8c801be1a8386d02992
BLAKE2b-256 4b0c5a19601bc630657d572f4af11e449a5d4757640e853022280d41c63b30b6

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page