Skip to main content

PowerMCP - MCP servers for power system software like PowerWorld and OpenDSS

Project description

PowerMCP ⚡

License: MIT Python Version

PowerMCP is an open-source collection of MCP servers for power system software like PowerWorld and OpenDSS. These tools enable LLMs to directly interact with power system applications, facilitating intelligent coordination, simulation, and control in the energy domain.

🌟 What is MCP?

The Model Context Protocol (MCP) is a revolutionary standard that enables AI applications to seamlessly connect with various external tools. Think of MCP as a universal adapter for AI applications, similar to what USB-C is for physical devices. It provides:

  • Standardized connections to power system software and data sources
  • Secure and efficient data exchange between AI agents and power systems
  • Reusable components for building intelligent power system applications
  • Interoperability between different AI models and power system tools

🤝 Our Community Vision

We're building an open-source community focused on accelerating AI adoption in the power domain through MCP. Our goals are:

  • Collaboration: Bring together power system experts, AI researchers, and software developers
  • Innovation: Create and share MCP servers for various power system software and tools
  • Education: Provide resources and examples for implementing AI in power systems
  • Standardization: Develop best practices for AI integration in the energy sector

🚀 Getting Started with MCP

📖 Quick Tutorial

🚀 New to PowerMCP? Start here!

📋 PowerMCP Tutorial PDF - Your complete guide to getting started with PowerMCP

This comprehensive tutorial will walk you through everything you need to know to begin using PowerMCP effectively.

Easy Fully Offline Quick Start - No API Keys

Using commericial AI models in Claude Desktop or Cursor.ai is great, but if your company security policies dictate, you can run these models fully offline and keep your confidential power flow information private, away from prying eyes.

Ollama setup for local AI model
  1. Install Ollama from https://ollama.com/download/windows

  2. From the model dropdown box, download a tool-capable AI model like GPT-oss or qwen3

You might have trouble downloading the models from online through Ollama. You can copy the models from another machine if you zip your %USERPROFILE%/.ollama/models folder and bring it from a machine that has network access to HuggingFace.

  1. Serve the model by enabling the option in Ollama settings or by running ollama serve
MCPHost for local MCP protocol handling
  1. Install the GO programming language from https://go.dev/dl

  2. Clone the MCPHost program from Github using

go install github.com/mark3labs/mcphost@latest
  1. Setup your config.json file Open your config.json file in a text editor. In the JSON list of tools, add the Powerflow programs you have installed on your computer. For example, PSLF
{
  "mcpServers": {
    "pslf": {
      "command": "python",
      "args": ["PSLF/pslf_mcp.py"]
    }
  }
}

Or for PowerWorld.

{
  "mcpServers": {
    "powerworld": {
      "command": "python",
      "args": ["PowerWorld/powerworld_mcp.py"]
    }
  }
}
  1. Start the MCP server, replacing the model name and config file with your preferred option.
mcphost -m ollama:qwen3:4b --config .\config.json

Video Demos

Check out these demos showcasing PowerMCP in action:

  • Contingency Evaluation Demo: An LLM automatically operates power system software, such as PowerWorld and pandapower, to perform contingency analysis and generate professional reports.

  • Loadgrowth Evaluation Demo: An LLM automatically operates power system software, such as PowerWorld, to evaluate different load growth scenarios and generate professional reports with recommendations.

Useful MCP Tutorials

MCP follows a client-server architecture where:

  • Hosts are LLM applications (like Claude Desktop or IDEs) that initiate connections
  • Clients maintain 1:1 connections with servers, inside the host application
  • Servers provide context, tools, and prompts to clients

Check out these helpful tutorials to get started with MCP:

Using with LLMs

To use these MCP tools with an LLM:

  1. Install the MCP Python SDK:
pip install mcp-server-git
  1. Run your MCP server:
python your_server.py
  1. Configure your LLM application (e.g., Claude Desktop, Cursor) to use the MCP server:
{
  "mcpServers": {
    "servername": {
      "command": "python",
      "args": ["your_server.py"]
    }
  }
}

For instance, for pandapower you could configure the server as follows:

{
  "mcpServers": {
    "pandapower": {
      "command": "python",
      "args": ["pandapower/panda_mcp.py"]
    }
  }
}

📚 Documentation

For detailed documentation about MCP, please visit:

🤝 Contributing

We welcome contributions! Please see our Contributing Guidelines for details.

📄 License

This project is licensed under the MIT License - see the LICENSE file for details.

🙏 Acknowledgments

Core Team

Special Thanks

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

iflow_mcp_power_agent_powermcp-0.1.0.tar.gz (16.2 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

File details

Details for the file iflow_mcp_power_agent_powermcp-0.1.0.tar.gz.

File metadata

  • Download URL: iflow_mcp_power_agent_powermcp-0.1.0.tar.gz
  • Upload date:
  • Size: 16.2 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: uv/0.10.0 {"installer":{"name":"uv","version":"0.10.0","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Debian GNU/Linux","version":"13","id":"trixie","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":null}

File hashes

Hashes for iflow_mcp_power_agent_powermcp-0.1.0.tar.gz
Algorithm Hash digest
SHA256 7f12cbb9bff1fe232cec88d1ebb328886ad462765869185a82c735f38a8d8de4
MD5 30bda3e127f2dd509f76b2b65be264e3
BLAKE2b-256 a03ce81954a97041c4a126d253f20586c628cc38d733862646885a52d5a8abe8

See more details on using hashes here.

File details

Details for the file iflow_mcp_power_agent_powermcp-0.1.0-py3-none-any.whl.

File metadata

  • Download URL: iflow_mcp_power_agent_powermcp-0.1.0-py3-none-any.whl
  • Upload date:
  • Size: 18.0 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: uv/0.10.0 {"installer":{"name":"uv","version":"0.10.0","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Debian GNU/Linux","version":"13","id":"trixie","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":null}

File hashes

Hashes for iflow_mcp_power_agent_powermcp-0.1.0-py3-none-any.whl
Algorithm Hash digest
SHA256 cebb64965428e48082c72f19cbe323e10b4c893f264d07d99b56520eedd2a40e
MD5 cb9084e2d7f72254541b22789168667b
BLAKE2b-256 d64ae4a9d632d56c56b172e5391d7a8e2e96512d4585247c8e392de2d847b7e2

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page