Skip to main content

Colab MCP - Local stdio MCP server to expose Claude/IDE logs as tools/resources

Project description

Colab MCP 🔗

Stop losing context when you switch between AI coding tools.

A Model Context Protocol (MCP) server that lets Claude Code, Cursor, Codex, and other AI coding assistants share logs and session history with each other.

PyPI version License: MIT

The Problem

You're coding with Claude Code. You make progress. Then you switch to Cursor to test something. Now you've lost all your context. You explain everything again. Then you jump to Codex. Explain it all over again.

It's exhausting.

The Solution

Colab MCP is a shared MCP server that exposes your chat logs, terminal history, and IDE events as tools and resources across all your AI coding assistants.

When you switch tools, your AI already knows what you were working on. No more copy-pasting. No more re-explaining. Just continuous flow.


✨ Features

  • 🔄 Share context across tools - Claude Code, Cursor, Codex, Gemini
  • 📜 Access chat transcripts from previous sessions
  • 🔍 Search across all logs - find that conversation from last week
  • 🎯 Session summaries - quick overview of what you were working on
  • 🖥️ Terminal & IDE event tracking - see what commands were run
  • 🚀 Fast setup - one command to install across all your tools

🏗️ How It Works

Colab MCP sits between your AI coding tools and your local log files, giving all your assistants access to shared context.

graph LR
    A[AI Tools] --> B[Colab MCP]
    B --> C[Your Logs]
    
    style B fill:#e8f4f8,stroke:#4a90a4,stroke-width:2px
    style A fill:#f5f5f5,stroke:#888,stroke-width:1px
    style C fill:#f5f5f5,stroke:#888,stroke-width:1px

The Flow:

  1. You work with Claude Code - Have a great conversation about architecture
  2. Logs are saved automatically - Claude stores the session in ~/.claude/
  3. You switch to Cursor - Time to write some code
  4. Cursor asks Colab MCP - "What was discussed earlier?"
  5. MCP reads the logs - Fetches your Claude session from disk
  6. Context restored - Cursor now knows everything Claude knew

No cloud sync. No APIs. Just local files read by a local server.

Installation is simple:

  • Run sudo ./install.py
  • Installer detects which AI tools you have (Claude, Cursor, Codex, Gemini)
  • Choose which ones to configure
  • Installer writes MCP config files for each tool
  • Restart your tools - done!

Context sharing in practice:

You can ask any AI tool things like:

  • "What was I working on yesterday?"
  • "Search my logs for authentication discussions"
  • "Summarize my last Cursor session"
  • "What errors did I hit this morning?"

And they'll actually know, because they can all read the same logs through Colab MCP.


🚀 Quick Start

1. Install

pip install colab-mcp

2. Configure Your AI Tools

Run the interactive installer:

sudo python -m colab_mcp.install

Or use the standalone installer script:

sudo ./install.py

The installer will:

  • 🔍 Detect which AI coding tools you have installed
  • ✅ Let you choose which ones to configure
  • ⚙️ Add Colab MCP to their MCP server configs
  • 📝 Give you instructions to restart each tool

3. Restart Your AI Tools

Restart Claude Code, Cursor, Codex, or whichever tools you configured.

That's it! 🎉


📖 Usage

Once installed, Colab MCP exposes several tools and resources to your AI assistants:

Tools

  • list_sessions - Get a list of all coding sessions
  • fetch_transcript - Retrieve the full transcript of a session
  • summarize_session - Get a quick summary of what happened
  • search_logs - Search across all logs (chat, MCP, IDE events)
  • codex_status - Check recent Codex CLI activity

Example Prompts

Try asking your AI assistant:

"What was I working on in my last session?"

"Search my logs for discussions about authentication"

"Summarize my session from yesterday afternoon"

"What errors did I encounter in the last hour?"


🛠️ Manual Configuration

If you prefer to configure manually, add this to your MCP config:

Claude Code (~/.claude/mcp.json)

{
  "servers": {
    "colab-mcp": {
      "command": "colab-mcp",
      "env": {
        "CLAUDE_HOME": "/home/yourusername/.claude",
        "CURSOR_LOGS": "/home/yourusername/.cursor-server/data/logs",
        "TMPDIR": "/tmp"
      }
    }
  }
}

Cursor (~/.cursor/mcp.json)

{
  "mcpServers": {
    "colab-mcp": {
      "command": "colab-mcp",
      "env": {
        "CLAUDE_HOME": "/home/yourusername/.claude",
        "CURSOR_LOGS": "/home/yourusername/.cursor-server/data/logs",
        "TMPDIR": "/tmp"
      }
    }
  }
}

Codex (~/.codex/config.toml)

[mcp_servers.colab-mcp]
command = "colab-mcp"
args = []
env = { CLAUDE_HOME = "/home/yourusername/.claude", CURSOR_LOGS = "/home/yourusername/.cursor-server/data/logs", TMPDIR = "/tmp" }

🗂️ Architecture

graph TB
    subgraph AI["AI Tools"]
        Claude[Claude Code]
        Cursor[Cursor]
        Codex[Codex]
    end
    
    MCP[Colab MCP Server]
    
    subgraph Logs["Log Files"]
        Chat[Chat History]
        IDE[IDE Events]
        Term[Terminal]
    end
    
    Claude --> MCP
    Cursor --> MCP
    Codex --> MCP
    
    MCP --> Chat
    MCP --> IDE
    MCP --> Term
    
    style MCP fill:#e8f4f8,stroke:#4a90a4,stroke-width:2px
    style AI fill:#f9f9f9,stroke:#ccc
    style Logs fill:#f9f9f9,stroke:#ccc

🤝 Contributing

Contributions are welcome! Check out the docs/ folder for more detailed information about how Colab MCP works.


📝 License

MIT License - see LICENSE for details.


🙏 Acknowledgments

Built with FastMCP - the fastest way to build MCP servers in Python.


Made with ❤️ by developers tired of losing context

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

colab_mcp-0.1.0.tar.gz (28.6 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

colab_mcp-0.1.0-py3-none-any.whl (11.9 kB view details)

Uploaded Python 3

File details

Details for the file colab_mcp-0.1.0.tar.gz.

File metadata

  • Download URL: colab_mcp-0.1.0.tar.gz
  • Upload date:
  • Size: 28.6 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for colab_mcp-0.1.0.tar.gz
Algorithm Hash digest
SHA256 b462e87d8745d37523f67baeaeae8748c84b1040204ddef5f0e341b932f0a9be
MD5 5ed61f89f44d235a2e891c48b265f469
BLAKE2b-256 04ac798e970149f74041f17d63db09cc54c7b6070c3d4590098002b4a765abe7

See more details on using hashes here.

File details

Details for the file colab_mcp-0.1.0-py3-none-any.whl.

File metadata

  • Download URL: colab_mcp-0.1.0-py3-none-any.whl
  • Upload date:
  • Size: 11.9 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for colab_mcp-0.1.0-py3-none-any.whl
Algorithm Hash digest
SHA256 0dae5bcc04e76505ca953a117949947c297af0f48ea2092a06d6f98d71dece8b
MD5 a54b3c6cc1109c7d11e4a683cbbbdd8a
BLAKE2b-256 c12489f4841d910020faeed869face90e3b541c6391483b0819bfa60b3a6fa7b

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page