AI Ops - Advanced artificial intelligence capabilities for Nautobot through multi-provider LLM architecture and Model Context Protocol
Project description
AI Ops
Overview
The AI Ops app brings advanced artificial intelligence capabilities to Nautobot through a flexible multi-provider architecture and the Model Context Protocol (MCP). This app provides an intelligent chat assistant that can interact with your Nautobot environment, external MCP servers, and other integrated systems to help automate operational tasks, answer questions, and provide insights based on your network infrastructure data.
At its core, AI Ops leverages LangGraph and LangChain to orchestrate conversations with Large Language Models (LLMs) from multiple providers (Ollama, OpenAI, Azure AI, Anthropic, HuggingFace, and custom providers), maintaining conversation context through checkpointed sessions stored in Redis. The app supports flexible LLM provider and model management, allowing administrators to define multiple providers and models for various use cases. A powerful middleware system enables request/response processing with features like caching, logging, validation, and retry logic. The multi-MCP server architecture enables the AI assistant to connect to both internal and external MCP servers, providing extensible tool access for network automation, data retrieval, and operational workflows. Production-ready features include automated health monitoring for MCP servers, middleware cache management, automatic status tracking, conversation persistence, and scheduled maintenance tasks to maintain optimal performance.
Note: This project is actively evolving. We're continuously adding new features, providers, and capabilities. Check the Release Notes for the latest updates and the GitHub Issues for upcoming features.
Key Features
- Multi-Provider LLM Support: Use models from Ollama (local), OpenAI, Azure AI, Anthropic, HuggingFace, or implement custom providers
- LLM Provider Management: Configure and manage multiple LLM providers with provider-specific settings and handler classes
- LLM Model Management: Configure multiple models from different providers with varying capabilities, temperature settings, and configurations
- Middleware System: Apply middleware chains to models for caching, logging, validation, retry logic, rate limiting, and custom processing
- Priority-Based Middleware Execution: Control middleware execution order (1-100) with pre and post-processing phases
- AI Chat Assistant: Interactive chat interface that understands and responds to natural language queries about your Nautobot environment
- MCP Server Integration: Connect to internal and external Model Context Protocol servers to extend capabilities with custom tools
- Automated Health Monitoring: Scheduled health checks for MCP servers with retry logic and automatic cache invalidation
- Conversation Persistence: Checkpoint-based conversation management using Redis ensures context is maintained across sessions
- Secure Configuration: API keys and credentials managed through Nautobot's Secret objects, never stored directly
- Scheduled Tasks: Background jobs for checkpoint cleanup, MCP server health monitoring, and middleware cache management
- RESTful API: Full API support for programmatic access to all models (providers, models, middleware, MCP servers)
- Environment-Aware: Supports LAB (local development with Ollama), NONPROD, and PROD environments
More screenshots and detailed use cases can be found in the Using the App page in the documentation.
Requirements
- Nautobot 2.4.22+
- Python 3.10 - 3.12
- Redis (for conversation checkpointing and caching)
- At least one LLM provider:
- Ollama (local, free): For development and testing
- OpenAI API: For OpenAI models (requires API key)
- Azure OpenAI: For Azure-hosted models (requires subscription)
- Anthropic API: For Claude models (requires API key)
- HuggingFace: For HuggingFace models (requires API key)
- Custom: Implement your own provider handler
- Optional: MCP servers for extended functionality
Documentation
Full documentation for this App can be found at kvncampos.github.io/nautobot-ai-ops:
- User Guide - Overview, Using the App, Getting Started.
- Administrator Guide - How to Install, Configure, Upgrade, or Uninstall the App.
- Developer Guide - Extending the App, Code Reference, Contribution Guide.
- Release Notes / Changelog.
- Frequently Asked Questions.
Contributing to the Documentation
You can find all the Markdown source for the App documentation under the docs folder in this repository. For simple edits, a Markdown capable editor is sufficient: clone the repository and edit away.
If you need to view the fully-generated documentation site, you can build it with MkDocs. A container hosting the documentation can be started using the invoke commands (details in the Development Environment Guide) on http://localhost:8001. Using this container, as your changes to the documentation are saved, they will be automatically rebuilt and any pages currently being viewed will be reloaded in your browser.
Any PRs with fixes or improvements are very welcome!
Questions
For any questions or comments, please check the FAQ first. Feel free to also swing by the Network to Code Slack (channel #nautobot), sign up here if you don't have an account.
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file nautobot_ai_ops-1.0.3.tar.gz.
File metadata
- Download URL: nautobot_ai_ops-1.0.3.tar.gz
- Upload date:
- Size: 96.0 kB
- Tags: Source
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
722e86803cd9310f850964244111d1b8e9af1ba9a64b7d8ef2c2589e1d2a5779
|
|
| MD5 |
f950f977624b3793799b97c25b910885
|
|
| BLAKE2b-256 |
2f103d33859f8c63d2d2050b64686dfdde90f1da57c17b0d4f1cd5f81f59633d
|
Provenance
The following attestation bundles were made for nautobot_ai_ops-1.0.3.tar.gz:
Publisher:
publish-pypi.yml on kvncampos/nautobot-ai-ops
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
nautobot_ai_ops-1.0.3.tar.gz -
Subject digest:
722e86803cd9310f850964244111d1b8e9af1ba9a64b7d8ef2c2589e1d2a5779 - Sigstore transparency entry: 814660944
- Sigstore integration time:
-
Permalink:
kvncampos/nautobot-ai-ops@6db7f4f984b1df5b78489a4e6ce23f45a35bb0d0 -
Branch / Tag:
refs/tags/1.0.3 - Owner: https://github.com/kvncampos
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
publish-pypi.yml@6db7f4f984b1df5b78489a4e6ce23f45a35bb0d0 -
Trigger Event:
push
-
Statement type:
File details
Details for the file nautobot_ai_ops-1.0.3-py3-none-any.whl.
File metadata
- Download URL: nautobot_ai_ops-1.0.3-py3-none-any.whl
- Upload date:
- Size: 124.3 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
5e084f68514acd42373e5f68a6d71d9e1cedcbc4696681beb2e94dfad5e21133
|
|
| MD5 |
e58f4229a3380d2f0ad01c41c32159e3
|
|
| BLAKE2b-256 |
7fb4a7577bc86c49dcec3cb84681232ecaf9ae4a4d1cca06285f8a00decb50a6
|
Provenance
The following attestation bundles were made for nautobot_ai_ops-1.0.3-py3-none-any.whl:
Publisher:
publish-pypi.yml on kvncampos/nautobot-ai-ops
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
nautobot_ai_ops-1.0.3-py3-none-any.whl -
Subject digest:
5e084f68514acd42373e5f68a6d71d9e1cedcbc4696681beb2e94dfad5e21133 - Sigstore transparency entry: 814660945
- Sigstore integration time:
-
Permalink:
kvncampos/nautobot-ai-ops@6db7f4f984b1df5b78489a4e6ce23f45a35bb0d0 -
Branch / Tag:
refs/tags/1.0.3 - Owner: https://github.com/kvncampos
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
publish-pypi.yml@6db7f4f984b1df5b78489a4e6ce23f45a35bb0d0 -
Trigger Event:
push
-
Statement type: