Skip to main content

AI coding agent — multi-provider, privacy-first, trust layer

Project description

Rewriting Project Claw Code

⭐ The fastest repo in history to surpass 50K stars, reaching the milestone in just 2 hours after publication ⭐

Star History Chart

Claw

Better Harness Tools, not merely storing the archive of leaked Claude Code

Sponsor on GitHub

[!IMPORTANT] Rust port is now in progress on the dev/rust branch and is expected to be merged into main today. The Rust implementation aims to deliver a faster, memory-safe harness runtime. Stay tuned — this will be the definitive version of the project.

If you find this work useful, consider sponsoring @instructkr on GitHub to support continued open-source harness engineering research.


Backstory

At 4 AM on March 31, 2026, I woke up to my phone blowing up with notifications. The Claude Code source had been exposed, and the entire dev community was in a frenzy. My girlfriend in Korea was genuinely worried I might face legal action from Anthropic just for having the code on my machine — so I did what any engineer would do under pressure: I sat down, ported the core features to Python from scratch, and pushed it before the sun came up.

The whole thing was orchestrated end-to-end using oh-my-codex (OmX) by @bellman_ych — a workflow layer built on top of OpenAI's Codex (@OpenAIDevs). I used $team mode for parallel code review and $ralph mode for persistent execution loops with architect-level verification. The entire porting session — from reading the original harness structure to producing a working Python tree with tests — was driven through OmX orchestration.

The result is a clean-room Python rewrite that captures the architectural patterns of Claude Code's agent harness without copying any proprietary source. I'm now actively collaborating with @bellman_ych — the creator of OmX himself — to push this further. The basic Python foundation is already in place and functional, but we're just getting started. Stay tuned — a much more capable version is on the way.

https://github.com/instructkr/claw-code

Tweet screenshot

The Creators Featured in Wall Street Journal For Avid Claude Code Fans

I've been deeply interested in harness engineering — studying how agent systems wire tools, orchestrate tasks, and manage runtime context. This isn't a sudden thing. The Wall Street Journal featured my work earlier this month, documenting how I've been one of the most active power users exploring these systems:

AI startup worker Sigrid Jin, who attended the Seoul dinner, single-handedly used 25 billion of Claude Code tokens last year. At the time, usage limits were looser, allowing early enthusiasts to reach tens of billions of tokens at a very low cost.

Despite his countless hours with Claude Code, Jin isn't faithful to any one AI lab. The tools available have different strengths and weaknesses, he said. Codex is better at reasoning, while Claude Code generates cleaner, more shareable code.

Jin flew to San Francisco in February for Claude Code's first birthday party, where attendees waited in line to compare notes with Cherny. The crowd included a practicing cardiologist from Belgium who had built an app to help patients navigate care, and a California lawyer who made a tool for automating building permit approvals using Claude Code.

"It was basically like a sharing party," Jin said. "There were lawyers, there were doctors, there were dentists. They did not have software engineering backgrounds."

The Wall Street Journal, March 21, 2026, "The Trillion Dollar Race to Automate Our Entire Lives"

WSJ Feature


Porting Status

The main source tree is now Python-first.

  • src/ contains the active Python porting workspace
  • tests/ verifies the current Python workspace
  • the exposed snapshot is no longer part of the tracked repository state

The current Python workspace is not yet a complete one-to-one replacement for the original system, but the primary implementation surface is now Python.

Why this rewrite exists

I originally studied the exposed codebase to understand its harness, tool wiring, and agent workflow. After spending more time with the legal and ethical questions—and after reading the essay linked below—I did not want the exposed snapshot itself to remain the main tracked source tree.

This repository now focuses on Python porting work instead.

Repository Layout

.
├── src/                                # Python porting workspace
│   ├── __init__.py
│   ├── commands.py
│   ├── main.py
│   ├── models.py
│   ├── port_manifest.py
│   ├── query_engine.py
│   ├── task.py
│   └── tools.py
├── tests/                              # Python verification
├── assets/omx/                         # OmX workflow screenshots
├── 2026-03-09-is-legal-the-same-as-legitimate-ai-reimplementation-and-the-erosion-of-copyleft.md
└── README.md

Python Workspace Overview

The new Python src/ tree currently provides:

  • port_manifest.py — summarizes the current Python workspace structure
  • models.py — dataclasses for subsystems, modules, and backlog state
  • commands.py — Python-side command port metadata
  • tools.py — Python-side tool port metadata
  • query_engine.py — renders a Python porting summary from the active workspace
  • main.py — a CLI entrypoint for manifest and summary output

Quickstart

Render the Python porting summary:

python3 -m src.main summary

Print the current Python workspace manifest:

python3 -m src.main manifest

List the current Python modules:

python3 -m src.main subsystems --limit 16

Run verification:

python3 -m unittest discover -s tests -v

Run the parity audit against the local ignored archive (when present):

python3 -m src.main parity-audit

Inspect mirrored command/tool inventories:

python3 -m src.main commands --limit 10
python3 -m src.main tools --limit 10

Current Parity Checkpoint

The port now mirrors the archived root-entry file surface, top-level subsystem names, and command/tool inventories much more closely than before. However, it is not yet a full runtime-equivalent replacement for the original TypeScript system; the Python tree still contains fewer executable runtime slices than the archived source.

Built with oh-my-codex

The restructuring and documentation work on this repository was AI-assisted and orchestrated with Yeachan Heo's oh-my-codex (OmX), layered on top of Codex.

  • $team mode: used for coordinated parallel review and architectural feedback
  • $ralph mode: used for persistent execution, verification, and completion discipline
  • Codex-driven workflow: used to turn the main src/ tree into a Python-first porting workspace

OmX workflow screenshots

OmX workflow screenshot 1

Ralph/team orchestration view while the README and essay context were being reviewed in terminal panes.

OmX workflow screenshot 2

Split-pane review and verification flow during the final README wording pass.

Community

instructkr

Join the instructkr Discord — the best Korean language model community. Come chat about LLMs, harness engineering, agent workflows, and everything in between.

Discord

Star History

See the chart at the top of this README.

Ownership / Affiliation Disclaimer

  • This repository does not claim ownership of the original Claude Code source material.
  • This repository is not affiliated with, endorsed by, or maintained by Anthropic.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

maestro_ai-0.9.0.tar.gz (6.3 MB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

maestro_ai-0.9.0-py3-none-any.whl (300.0 kB view details)

Uploaded Python 3

File details

Details for the file maestro_ai-0.9.0.tar.gz.

File metadata

  • Download URL: maestro_ai-0.9.0.tar.gz
  • Upload date:
  • Size: 6.3 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.11.8

File hashes

Hashes for maestro_ai-0.9.0.tar.gz
Algorithm Hash digest
SHA256 67281cfab1e7a0ca815c36fd8f694a5ce45c75822389975211c2e342fae48867
MD5 8b79e474e9a6dd9032e253345ef10fd5
BLAKE2b-256 274c8f92c5583c0829f4393481d41c3629317178054a1e119d34d40286a9072b

See more details on using hashes here.

File details

Details for the file maestro_ai-0.9.0-py3-none-any.whl.

File metadata

  • Download URL: maestro_ai-0.9.0-py3-none-any.whl
  • Upload date:
  • Size: 300.0 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.11.8

File hashes

Hashes for maestro_ai-0.9.0-py3-none-any.whl
Algorithm Hash digest
SHA256 e04013f6ba1936e612712dc0b42ef8e67f5aefb83ca7f1f39f85d8cfe7d4a8c4
MD5 e5b04e6c65193e4e1b8020caf933c5df
BLAKE2b-256 014699daa64ff5d7a8b2f4138f712fe1ad638b473af2cd8e0d7211c08620ad65

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page