Skip to main content

Managing a unified ecosystem of LLM providers, agents, and MCP servers.

Project description

🤖 AgentManager: Managing a unified ecosystem of LLM providers, agents, and MCP servers.

AgentManager is a Python package that provides a unified, high-level system for working with Large Language Models (LLMs). It simplifies managing provider credentials, selecting models, and constructing AI agents. It also supports integrating external tools via the Model Context Protocol (MCP) for more advanced workflows.


✨ Features

  • Universal Provider Support: Seamlessly connect to various cloud LLM providers (e.g., Google, OpenAI, Ollama, Mistral, Groq) through a single line code.
  • Agent Construction: Quickly set up basic agents or advanced tool-powered agents with minimal setup.
  • MCP Integration: Easily incorporate tools from a MCP Server to enable complex, external actions.
  • Chat Support: Supports both single-turn interactions and stateful, continuous conversations.
  • UI (Optional): Interactive UI to explore the AgentManager.

📄 Requirements

  • Python 3.11+

🛠️ Installation

  • Install the core package without UI features:

    pip install agentmanager
    
  • To use the interactive UI, install with the [ui] extras:

    pip install agentmanager[ui]
    
  • If you already have an older version of AgentManager installed, upgrade to get the latest features:

    pip install --upgrade agentmanager[ui]
    

Launching the UI

You can dive right in and explore AgentManager’s features through the interactive UI by running the following command in your terminal:

agentmanager-ui

This opens a intaractive interface in your browser, where you can:

  • Manage LLM providers
  • Configure and run agents
  • Add and use multiple MCP tools

It’s a hands-on way to see everything in action.

AgentManager UI


ℹ️ Utility Methods

The CloudAgentManager provides helpful utility methods for discovering supported providers and models.

Method Description
cloud_agent_manager.get_providers() Returns a list of all supported LLM provider names.
cloud_agent_manager.get_models(provider_name) Returns a list of all available model names for a given provider.
cloud_agent_manager.get_provider_key(provider_name) Returns the URL link where you can obtain your API key for the provider.

Example Utility Usage

from agentmanager import CloudAgentManager

cloud_agent_manager = CloudAgentManager()

# Get all supported providers
providers = cloud_agent_manager.get_providers()
print("Supported Providers:", providers)
# Output might be: ['OpenAI', 'Google', 'Ollama', ...]

# Get models for a specific provider
google_models = cloud_agent_manager.get_models("ollama")
print("Ollama Models:", ollama_models)
    
# Get API key page link
page_link_for_api_key = cloud_agent_manager.get_provider_key("mistral")
print("Mistral Key Link:", page_link_for_api_key)

🚀 Quick Start

The core functionality is encapsulated in the CloudAgentManager class. Here is a basic example for a single-turn chat.

Single-Turn Chat Example

This script demonstrates initializing an LLM, creating a basic agent, and getting a single response.

import asyncio
from agentmanager import CloudAgentManager

# The following constants would typically be read from environment variables or a secure vault
# Placeholder values are used for demonstration.
PROVIDER = "google"
API_KEY = "YOUR_API_KEY_HERE" # !!! REPLACE WITH YOUR ACTUAL API KEY !!!
MODEL_NAME = "gemini-2.5-flash"

async def single_chat():
    # 1️⃣ Initialize the CloudAgentManager
    cloud_agent_manager = CloudAgentManager()

    # 2️⃣ Prepare the LLM (handles API key and model validation)
    llm = cloud_agent_manager.prepare_llm(PROVIDER, API_KEY, MODEL_NAME)
        
    # 3️⃣ Prepare the agent (no MCP tools in this example)
    agent, tools = await cloud_agent_manager.prepare_agent(llm)
    
    # 4️⃣ Send message
    user_message = "what is the capital of India?"    
    response_messages = await cloud_agent_manager.get_agent_response(agent, user_message)
    
    for m in response_messages:
        print(m.content)


if __name__ == "__main__":
    asyncio.run(single_chat())

💬 Continuous Chat Loop

For an interactive, multi-turn conversation that maintains context using chat_history.

import asyncio
from agentmanager import CloudAgentManager
from typing import List

# The following constants would typically be read from environment variables or a secure vault
# Placeholder values are used for demonstration.
PROVIDER = "google"
API_KEY = "YOUR_API_KEY_HERE" # !!! REPLACE WITH YOUR ACTUAL API KEY !!!
MODEL_NAME = "gemini-2.5-flash"

async def chat_loop():
    # 1️⃣ Initialize the CloudAgentManager
    cloud_agent_manager = CloudAgentManager()

    # 2️⃣ Prepare LLM
    llm = cloud_agent_manager.prepare_llm(PROVIDER, API_KEY, MODEL_NAME)
        
    # 3️⃣ Prepare agent (e.g., without MCP tools)
    agent, tools = await cloud_agent_manager.prepare_agent(llm)

    # 4️⃣ Initialize chat history
    # The agent will use this list to maintain context across turns.
    chat_history: List[Any] = []

    # 5️⃣ Terminal loop
    print("\n--- Start Chat ---")
    print("Type 'exit' or 'quit' to end the session.")
    while True:
        user_input = input("You: ").strip()
        
        if user_input.lower() in {"exit", "quit"}:
            print("👋 Goodbye!")
            break

        if not user_input:
            continue

        try:
            # The cloud_agent_manager updates chat_history in place
            new_messages = await cloud_agent_manager.get_agent_response(agent, user_input, chat_history)
            
            for m in new_messages:
                print(f"Agent: {m.content}")
                
        except Exception as e:
            print(f"❌ Agent failed to respond: {e}")

if __name__ == "__main__":
    asyncio.run(chat_loop())

⚙️ Advanced: Agent with MCP Tools

If your agent needs to interact with external tools via a Model Context Protocol (MCP), you can provide a list of MCP server configurations during agent preparation.

Each MCP entry can include:

  • name (str): A unique name for the MCP server (required)
  • url (str): Your MCP server's URL (required)
  • headers (Optional[Dict[str, str]]): Dictionary of headers, e.g.
    {"Authorization": "Bearer XYZ", "X-Custom": "ABC123", ...}
    
import asyncio
from agentmanager import CloudAgentManager

# The following constants would typically be read from environment variables or a secure vault
# Placeholder values are used for demonstration.
PROVIDER = "google" # !!! REPLACE WITH PROVIDER YOU WANT TO USE !!!
API_KEY = "YOUR_API_KEY_HERE" # !!! REPLACE WITH PROVIDER'S API KEY !!!
MODEL_NAME = "gemini-2.5-flash" # !!! REPLACE WITH MODEL YOU WANT TO USE !!!

# Define multiple MCP configurations (name, url, optional multiple headers)
mcps=[
    {
        "name": "MyFirstServer",
        "url": "MCP_URL_1" # !!! REPLACE WITH YOUR MCP URL !!!
    },
    {
        "name": "MySecondServer",
        "url": "MCP_URL_2", # !!! REPLACE WITH YOUR MCP URL !!!
        # Optional: Custom headers for authentication/routing
        "header": {
            # !!! REPLACE WITH YOUR ACTUAL HEADER NAMES & VALUES !!!
            "FIRST_HEADER_NAME": "FIRST_HEADER_VALUE",
            "SECOND_HEADER_NAME": "SECOND_HEADER_VALUE",
            # Add more headers as needed ...
        }
    },
    # Add more MCP servers as needed ...
]

async def mcp_agent_example():

    # 1️⃣ Initialize the CloudAgentManager
    cloud_agent_manager = CloudAgentManager()

    # 2️⃣ Prepare LLM
    llm = cloud_agent_manager.prepare_llm(PROVIDER, API_KEY, MODEL_NAME)
    
    # 3️⃣ Prepare agent with MCP configuration
    agent, tools = await cloud_agent_manager.prepare_agent(llm, mcps)

    # 4️⃣ Initialize chat history
    # The agent will use this list to maintain context across turns.
    chat_history: List[Any] = []

    # 5️⃣ Terminal loop
    print("\n--- Start Chat ---")
    print("Type 'exit' or 'quit' to end the session.")
    while True:
        user_input = input("You: ").strip()
        
        if user_input.lower() in {"exit", "quit"}:
            print("👋 Goodbye!")
            break

        if not user_input:
            continue

        try:
            # The cloud_agent_manager updates chat_history in place
            new_messages = await cloud_agent_manager.get_agent_response(agent, user_input, chat_history)
            
            for m in new_messages:
                print(f"Agent: {m.content}")
                
        except Exception as e:
            print(f"❌ Agent failed to respond: {e}")

if __name__ == "__main__":
    asyncio.run(mcp_agent_example())

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

agentmanager-0.1.1.tar.gz (11.8 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

agentmanager-0.1.1-py3-none-any.whl (13.0 kB view details)

Uploaded Python 3

File details

Details for the file agentmanager-0.1.1.tar.gz.

File metadata

  • Download URL: agentmanager-0.1.1.tar.gz
  • Upload date:
  • Size: 11.8 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.9.7

File hashes

Hashes for agentmanager-0.1.1.tar.gz
Algorithm Hash digest
SHA256 c96cb42bf0bc12a5e553d8afc198ba81ae468769ccc163aca14de308bef4fd1f
MD5 59e93e94e016b06c6f67016ae4afc33a
BLAKE2b-256 d292b78f5c0aeefccad15837c6eef272c1df890a692e5bd1419b015875784903

See more details on using hashes here.

File details

Details for the file agentmanager-0.1.1-py3-none-any.whl.

File metadata

  • Download URL: agentmanager-0.1.1-py3-none-any.whl
  • Upload date:
  • Size: 13.0 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.9.7

File hashes

Hashes for agentmanager-0.1.1-py3-none-any.whl
Algorithm Hash digest
SHA256 fb750f3b9e2fc4c87e500ad69ca24bfdc1bbfb0ceab4cec5875144412617aa52
MD5 bb8387fe66c1c1011a6a8006b63bf007
BLAKE2b-256 229b1ad33a47c92b94f7895c525c1d57363bded887cbf7ea2e25ec20a8ab13c9

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page