Skip to main content

The simplest way to context engineer. Minimal streaming CLI clients for Claude and Gemini.

Project description

Raw LLM

The simplest way to context engineer.

Minimal, streaming CLI clients for Claude and Gemini that keep your conversations in plain JSON files.

PyPI version Python versions License: MIT

What is this?

Raw LLM is a pair of thin Python scripts that talk to the Anthropic and Google GenAI APIs. No frameworks, no agents, no abstractions you don't need. Just a prompt, a streaming response, and a JSON file you can version, diff, edit, and pipe.

The entire idea: your conversation is a file. You build context by editing that file. That's it. That's the context engineering.

Features

  • Streaming output — responses print token-by-token as they arrive
  • Conversation persistence — every exchange is saved to a plain JSON file you own
  • Resume any conversation — pass the JSON file back in to continue where you left off
  • Pipe-friendly — reads from stdin, writes content to stdout, writes diagnostics to stderr
  • Colored output — reasoning in gray (stderr), content in cyan (stdout), auto-disabled when piped
  • Conflict detection — refuses to overwrite a conversation file modified by another process
  • Symlink to switch models — symlink claude.py as opus or haiku to change the default model

Installation

From PyPI

pip install raw-llm

This installs the claude, sonnet, opus, haiku, and gemini commands globally.

From source

git clone https://github.com/rodolfovillaruz/raw-llm.git
cd raw-llm
pip install .

Development install

git clone https://github.com/rodolfovillaruz/raw-llm.git
cd raw-llm
pip install -e ".[dev]"

Set your API keys:

export ANTHROPIC_API_KEY="sk-ant-..."
export GEMINI_API_KEY="..."       # or GOOGLE_API_KEY, per google-genai docs

Usage

Start a new conversation

claude
# Type your prompt, then press Ctrl+D to submit
echo "Explain monads in one paragraph" | claude
gemini

Resume an existing conversation

claude .prompt/some-conversation.json

The JSON file contains the full message history. Edit it with any text editor to reshape context before your next turn.

Pipe a file as context

cat code.py | claude conversation.json

Switch models

# By flag
claude -m claude-opus-4-6

# By command name
opus
haiku
sonnet
Command Default model
claude / sonnet claude-sonnet-4-6
opus claude-opus-4-6
haiku claude-haiku-4-5
gemini gemini-3.1-pro-preview

Options

usage: claude [-h] [-n] [-v] [-m MODEL] [-t MAX_TOKENS] [-i] [conversation_file]

positional arguments:
  conversation_file         JSON file to resume (omit to start fresh)

options:
  -n, --dry-run             Build the prompt but don't send it
  -v, --verbose             Show model name and prompt preview
  -m, --model MODEL         Override the default model
  -t, --max-tokens TOKENS   Cap the response length
  -i, --interactive         Interactive REPL mode

Conversation format

Conversations are stored as a JSON array of message objects, the same shape both APIs understand:

[
  {
    "role": "user",
    "content": "What is context engineering?"
  },
  {
    "role": "assistant",
    "content": "Context engineering is the practice of ..."
  }
]

You can create these files by hand, merge them, truncate them, or generate them with other tools. Raw LLM doesn't care. It reads the array, appends your new message, streams the response, and appends that too.

Project structure

.
├── src/
│   └── raw_llm/
│       ├── claude.py       # Claude CLI client
│       ├── gemini.py       # Gemini CLI client
│       └── common.py       # Shared utilities (streaming, I/O, conversation management)
├── pyproject.toml          # Package configuration and entry points
├── Makefile                # Formatting, linting, typing
└── .prompt/                # Default directory for conversation files (auto-used if present)

Development

make fmt      # Format with black/isort
make lint     # Lint with pylint/flake8
make type     # Type-check with mypy
make all      # All of the above

Why?

Most LLM tools add layers between you and the model. Raw LLM removes them. The conversation is a file. The prompt is stdin. The response is stdout. Everything else is up to you.

License

MIT

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

raw_llm-1.0.4.tar.gz (12.5 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

raw_llm-1.0.4-py3-none-any.whl (11.1 kB view details)

Uploaded Python 3

File details

Details for the file raw_llm-1.0.4.tar.gz.

File metadata

  • Download URL: raw_llm-1.0.4.tar.gz
  • Upload date:
  • Size: 12.5 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.3

File hashes

Hashes for raw_llm-1.0.4.tar.gz
Algorithm Hash digest
SHA256 28bdf012f4d95b869dd3129929b84bb0f283a29c5738b1c673a26e3fc300f5e9
MD5 6d560293c5bad17cfdc66c13457e4ec9
BLAKE2b-256 62b0ce95fb55906103a87c4551c6fc5900d3b9d9c48f9f0abb8ea0769e2a2975

See more details on using hashes here.

File details

Details for the file raw_llm-1.0.4-py3-none-any.whl.

File metadata

  • Download URL: raw_llm-1.0.4-py3-none-any.whl
  • Upload date:
  • Size: 11.1 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.3

File hashes

Hashes for raw_llm-1.0.4-py3-none-any.whl
Algorithm Hash digest
SHA256 a258a68fd4a4c854f4032ae2b845978fc7dcf95a48c6bb52d6255596dffd924b
MD5 66e791961bbe300649b7696c6941fe84
BLAKE2b-256 6bcf70a2ae3f289844e5aed81a34b15b5a102887ca7fa59fc3b097079fe4cc68

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page