MCP server that turns any Senior-Junior workflow into an autonomous loop with a human decision maker. Seamless integration with Cursor, Claude Code, Codex, and any MCP client.
Project description
Autonomous Lab
MCP server that turns any senior-junior workflow into an autonomous loop. AI handles the execution. You make the decisions.
Vision
The bottleneck in knowledge work has never been execution. It is judgment -- knowing which questions matter, which results are meaningful, which directions to pursue. The people best equipped to make those calls spend most of their time on tasks that don't require their specific expertise.
Autonomous Lab shifts the hierarchy up by one level. AI agents assume the working roles -- principal investigator and trainee, tech lead and developer, attending and resident -- running the full design-execute-review loop. The human moves into the editorial position: the one who curates, judges, and steers. Your taste and judgment, rather than your labor, become the primary input.
This is not a copilot. It is a reorganization of the work unit itself.
Why this exists
Autonomous Lab is an MCP server. It runs inside the coding agent you already pay for -- Cursor, Claude Code, Windsurf, Codex CLI, or any MCP-compatible client. That means:
- No API key required. You don't need an OpenAI/Anthropic/Google key. The intelligence comes from whichever model your coding tool already uses.
- No extra cost. Your existing Cursor Pro, Claude Max, Windsurf, or Codex subscription is all you need. You are reusing an investment you have already made.
- No new app to learn. It plugs into your current workflow as a set of MCP tools.
Install
The easiest way: copy this page link into Claude Code, Cursor, or any coding agent and ask it to install Autonomous Lab for you. It will handle everything.
Or do it manually:
Add to your MCP client config (e.g. Cursor ~/.cursor/mcp.json):
{
"mcpServers": {
"autonomous-lab": {
"command": "uvx",
"args": ["autonomous-lab"],
"timeout": 600,
"env": {
"MCP_WEB_PORT": "8766"
}
}
}
}
Or if you installed via uv pip install:
{
"mcpServers": {
"autonomous-lab": {
"command": "autonomous-lab",
"timeout": 600,
"env": {
"MCP_WEB_PORT": "8766"
}
}
}
}
Then tell your agent: "Initialize an autonomous lab project on [your topic]."
What it does
Two AI personas (senior + junior) iterate on your project in a loop. They design, execute, write, and revise. You sit above them as the decision maker: editor, code reviewer, creative director, or whatever the domain calls for.
The loop:
autolab_next → (AI acts as role) → autolab_record → lab_meeting → autolab_next → ...
When work is ready, you review it. Accept, request revisions, or reject. The loop continues until you're satisfied.
Anatomy of the monitoring interface and editorial workflow. Top: the research loop (characters, meeting log, inventory, marketplace). Bottom: the editorial office (reviewer selection, reports, decision).
Key capabilities
- Zero additional cost: runs on your existing coding agent subscription. No separate API keys, no usage-based billing, no new accounts.
- Multi-agent orchestration: opt-in mode where each role (PI, Trainee, Reviewer) runs as a separate agent with its own context window. Uses your existing CLI subscriptions (Claude Code, Codex CLI, Cursor) -- no API keys needed. Falls back to single-agent on any failure.
- Skill containers: configure characters with any combination of SKILL.md files you already have. A PI with
scanpy + scientific-writing + statistical-analysisskills behaves differently from a Tech Lead withreact + typescript + code-reviewskills. - 24-hour sessions: the loop runs indefinitely. No timeout, no context loss. Sessions persist across disconnects with
autolab_resume. - Fully configurable: YAML character profiles control personality, expertise, goals, and available tools. Swap them in seconds.
- Domain-agnostic: research, software, consulting, legal, medical, creative, or anything with a senior-junior structure.
- Expert consultation: invite domain specialists mid-session for one-off advice without breaking the loop.
- Verified citations: built-in CrossRef integration for real, validated references (no hallucinated papers).
- Game-style monitoring UI: browser dashboard shows live progress, iteration history, and editorial controls.
MCP tools
| Tool | What it does |
|---|---|
autolab_init |
Initialize a new project |
autolab_resume |
Resume an interrupted session |
autolab_next |
Get the next role prompt (PI or Trainee) |
autolab_record |
Record a completed turn |
autolab_status |
Check project state |
autolab_cite |
Search, validate, and format citations |
autolab_consult |
Invite a domain expert |
autolab_editorial |
Wait for editor decision |
autolab_editor_act |
Execute editorial decision (AI fallback) |
autolab_create_character |
Build a character profile |
lab_meeting |
Pause for user feedback between turns |
Character example
name: Dr. Maria Chen
role: pi
title: Computational Biology PI
expertise: single-cell genomics, machine learning
goal: discover cell-type-specific regulatory programs
skills:
- scanpy
- scvi-tools
- scientific-writing
- statistical-analysis
personality:
- "Visionary: spots novel research directions"
- "Rigorous: demands statistical reproducibility"
Multi-agent mode
By default, a single AI agent plays all roles in sequence. Multi-agent mode spawns a dedicated agent for each role, giving each a fresh context window. Enable it in .autolab/config.yaml:
orchestration: multi
agents:
pi:
provider: claude-cli # Uses your Claude Code subscription
model: opus
trainee:
provider: claude-cli
model: sonnet
Multiple trainees
A PI can have multiple trainees working on different tasks in parallel. Each trainee gets its own context window and focus area:
orchestration: multi
agents:
pi:
provider: claude-cli
model: opus
trainees: # plural key
- name: "Data Analyst"
provider: claude-cli
model: sonnet
focus: "data analysis, statistics, figure generation"
- name: "Writer"
provider: codex-cli
model: o3
focus: "paper writing, LaTeX, literature review"
- name: "Code Developer"
provider: cursor-cli
model: sonnet-4
focus: "script development, testing, reproducibility"
All trainees run in parallel via asyncio.gather. Each sees a scoped prompt with their focus area and coordination notes about what the other trainees are handling. If one fails, the others continue.
Agent teams (Claude Code)
When running inside Claude Code with agent teams enabled (CLAUDE_CODE_EXPERIMENTAL_AGENT_TEAMS=1), multi-agent mode automatically upgrades to native agent teams. The PI becomes the team lead in delegate mode, trainees become teammates with their own context windows, and they coordinate via shared task lists and direct messaging. No extra config needed -- the same orchestration: multi config works for both modes.
Optional team-specific settings:
team:
delegate_mode: true # PI only coordinates, doesn't implement
require_plan_approval: false # Require PI approval before trainees act
Supported providers
| Provider | Cost | Requires |
|---|---|---|
claude-cli |
Free (Pro/Max sub) | Claude Code CLI installed |
codex-cli |
Free (Plus/Pro sub) | Codex CLI installed |
cursor-cli |
Free (Pro/Biz sub) | Cursor Agent CLI installed |
anthropic |
Pay-per-token | ANTHROPIC_API_KEY env var |
openai |
Pay-per-token | OPENAI_API_KEY env var |
CLI providers are the primary path -- they need zero API key configuration. If the CLI binary is not found or multi-agent fails for any reason, it falls back to single-agent mode automatically.
For API providers, install the optional dependencies:
pip install 'autonomous-lab[multi-agent]'
Remote / SSH environments
The monitoring web UI binds to 127.0.0.1 by default (local only). On a remote server, SSH session, or container, the UI will attempt to auto-detect and bind to 0.0.0.0 instead. If auto-detection doesn't match your setup, use one of the methods below.
Method 1: Environment variable (recommended)
Set MCP_WEB_HOST to 0.0.0.0 in your MCP config:
{
"mcpServers": {
"autonomous-lab": {
"command": "uvx",
"args": ["autonomous-lab"],
"timeout": 600,
"env": {
"MCP_WEB_HOST": "0.0.0.0",
"MCP_WEB_PORT": "8766"
}
}
}
}
Then open http://<remote-host-ip>:8766/lab in your local browser.
Method 2: SSH port forwarding
Keep the default config (127.0.0.1) and forward the port:
ssh -L 8766:localhost:8766 user@remote-host
Then open http://localhost:8766/lab locally.
| Variable | Purpose | Default |
|---|---|---|
MCP_WEB_HOST |
Bind address | auto-detected (0.0.0.0 if SSH/container, else 127.0.0.1) |
MCP_WEB_PORT |
Web UI port | 8765 |
Requirements
- Python >= 3.11
- An MCP-compatible client (Cursor, Claude Code, Codex CLI, Windsurf, etc.)
Acknowledgments
Autonomous Lab builds on these open-source projects:
- The Virtual Lab by James Zou Lab, Stanford (MIT) -- the concept of LLM agents as PI and scientists iterating through structured research meetings (Swanson et al., Nature 2025)
- mcp-feedback-enhanced by Minidoracat (MIT) -- Web UI, feedback loop, session management, and i18n infrastructure
- interactive-feedback-mcp by Fábio Ferreira (MIT) -- the original MCP feedback server
- biomni by Jure Leskovec Lab, Stanford (Apache 2.0) -- optional biomedical toolkit integration
License
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file autonomous_lab-0.7.2.tar.gz.
File metadata
- Download URL: autonomous_lab-0.7.2.tar.gz
- Upload date:
- Size: 5.9 MB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: uv/0.10.3 {"installer":{"name":"uv","version":"0.10.3","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"macOS","version":null,"id":null,"libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":null}
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
4b96ffef09542abf91bae83a3a4bdc238346e2ecf0187923a0ca24410b35d872
|
|
| MD5 |
d766343da6e304c2e198b523e2ecd783
|
|
| BLAKE2b-256 |
ae16073ef16aa627f82f66aacaf4a53911471f2d19b2c90e089b552f14d1f3cb
|
File details
Details for the file autonomous_lab-0.7.2-py3-none-any.whl.
File metadata
- Download URL: autonomous_lab-0.7.2-py3-none-any.whl
- Upload date:
- Size: 475.9 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: uv/0.10.3 {"installer":{"name":"uv","version":"0.10.3","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"macOS","version":null,"id":null,"libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":null}
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
fa9501daa99291165ab3786240e90aa6b4518e405e4a5ab29728e51628d6ada0
|
|
| MD5 |
f57c00b89cf1a4be9a7a353295496ff1
|
|
| BLAKE2b-256 |
200c32e3b66a6b98b25eda806d93347601b1d8c04de0e023c14d1f2d7980fdc6
|