Skip to main content

MCP servers for biological modelling: NeKo network analysis, MaBoSS Boolean dynamics, and PhysiCell multiscale simulation

Project description

MCP Bio‑Modelling Servers

This repository centralizes Model Context Protocol (MCP) servers that wrap Python‑based mechanistic / systems biology modelling tools. Each subfolder contains a server.py entrypoint plus a README describing the specific tool interface.

Current servers (see their own READMEs & upstream docs):

Tool Folder Upstream Documentation
MaBoSS MaBoSS/ https://github.com/colomoto/pyMaBoSS
NeKo NeKo/ https://github.com/sysbio-curie/Neko
PhysiCell (settings wrapper) PhysiCell/ https://github.com/marcorusc/PhysiCell_Settings

All servers are Python processes speaking MCP over stdio.


MCP Background

The Model Context Protocol standardizes how external tools expose tools and resources to AI assistants / IDEs. Spec & introduction: https://modelcontextprotocol.io/docs/getting-started/intro

Each server.py advertises modelling actions (e.g. run simulations, manage sessions) to any MCP‑aware client (e.g. VS Code with GitHub Copilot Chat MCP support).


Repository Layout

MaBoSS/    # MaBoSS MCP server (Boolean / stochastic models)
NeKo/      # NeKo MCP server
PhysiCell/ # PhysiCell settings / sessions MCP server
README.md

Consult the README within each tool folder for: purpose, required Python packages, and any model/data file expectations. Installation instructions for the modelling tools themselves live there (or in the upstream project links above) — they are intentionally not duplicated here.


Environment Assumption

All tools are Python‑based. Create (and manage) a single Conda environment that contains the dependencies for MaBoSS, NeKo, and PhysiCell. The exact creation commands are up to you (not prescribed here). Once created, note the absolute path to its Python interpreter (e.g. /home/you/miniforge3/envs/mcp_modelling/bin/python).


Configure in VS Code (GitHub Copilot Chat / MCP)

  1. Clone this repo somewhere stable (no spaces in path recommended).
  2. Open VS Code.
  3. Ensure the Copilot Chat (or other MCP‑capable) extension is installed and updated.
  4. Press Ctrl + Shift + P and search for the command that opens the MCP configuration JSON (e.g. “MCP: Open Configuration” or locate mcp.json). On Linux it typically lives at: ~/.config/Code/User/mcp.json.
  5. Add entries pointing to each server.py using the Conda environment’s Python.

Example (adapt paths to your system; based on the working setup):

{
  "servers": {
    "maboss": {
      "type": "stdio",
      "command": "/home/you/miniforge3/envs/mcp_modelling/bin/python",
      "args": [
        "/absolute/path/to/mcp-biomodelling-servers/MaBoSS/server.py"
      ],
      "env": {
        "PATH": "/home/you/miniforge3/envs/mcp_modelling/bin:${Path}",
        "CONDA_PREFIX": "/home/you/miniforge3/envs/mcp_modelling"
      }
    },
    "neko": {
      "type": "stdio",
      "command": "/home/you/miniforge3/envs/mcp_modelling/bin/python",
      "args": [
        "/absolute/path/to/mcp-biomodelling-servers/NeKo/server.py"
      ]
    },
    "physicell": {
      "type": "stdio",
      "command": "/home/you/miniforge3/envs/mcp_modelling/bin/python",
      "args": [
        "/absolute/path/to/mcp-biomodelling-servers/PhysiCell/server.py"
      ]
    }
  },
  "inputs": [
    {
      "id": "memory_file_path",
      "type": "promptString",
      "description": "Path to the memory storage file",
      "password": false
    }
  ]
}

Notes:

  • Replace /home/you/... and /absolute/path/to/... with your actual directories.
  • Keep all three servers referencing the same Conda interpreter to share installed libraries.
  • Add further environment variables (e.g. data directories) per tool README if required.

After saving, reload / restart VS Code so the MCP client reconnects.

Activation / usage guidance in VS Code: https://code.visualstudio.com/docs/copilot/chat/mcp-servers

You should then see the servers’ tools listed in the Copilot Chat “/tools” (or similar) UI. Invoke them by name with required parameters.


Adding Another Server

  1. Create a new folder with server.py and a README describing the underlying modelling tool and dependencies.
  2. Follow existing server structure for registering MCP tools.
  3. Update your mcp.json with a new block (use the same Conda Python path).
  4. Document any additional env vars in that folder README.

License

Project is MIT (see existing LICENSE file). Underlying tools retain their own licenses — consult upstream repositories.


Quick Reference

Action What to Do
Get tool install steps Open the tool’s subfolder README or upstream link
Ensure deps present Install into your chosen Conda env (user‑defined)
Configure MCP Edit ~/.config/Code/User/mcp.json as above
Reload servers Reload VS Code window
Learn MCP Spec: modelcontextprotocol.io; VS Code guide link above

Happy modelling!

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

mcp_biomodelling_servers-1.0.0.tar.gz (66.8 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

mcp_biomodelling_servers-1.0.0-py3-none-any.whl (72.7 kB view details)

Uploaded Python 3

File details

Details for the file mcp_biomodelling_servers-1.0.0.tar.gz.

File metadata

  • Download URL: mcp_biomodelling_servers-1.0.0.tar.gz
  • Upload date:
  • Size: 66.8 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.10.14

File hashes

Hashes for mcp_biomodelling_servers-1.0.0.tar.gz
Algorithm Hash digest
SHA256 52d43d1bc5809d371ddbe03573a74c90a11a1c94e599ccfe98f48feb0c002710
MD5 af121e1714382af700d3eacd35c61b5f
BLAKE2b-256 b251cfe95042728974d47914da3de962863e7c1802266b4de54d321c60fc4e58

See more details on using hashes here.

File details

Details for the file mcp_biomodelling_servers-1.0.0-py3-none-any.whl.

File metadata

File hashes

Hashes for mcp_biomodelling_servers-1.0.0-py3-none-any.whl
Algorithm Hash digest
SHA256 c75a9b3975cbfbbf086b0dff0febf700853bd6e70d8c29663850e38e073ef23b
MD5 43708e5bceefa5bbeeb4c89379aad219
BLAKE2b-256 1ecff4d87d680f03fc2d2a455f7c07af3fcc9cc2a4f29b5be3c5e0a7b1d302b8

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page