Skip to main content

Production agent framework on LangChain/LangGraph: nine execution patterns, persistent memory, skills, feedback, multi-level HITL, MCP, AGP protocol, runtime bridge, and observability hooks.

Project description

agloom

agloom

Build agents that route themselves.
One familiar API — classification, memory, streaming, guardrails, and learning included.

Nine execution patterns. Auto-selected per task. Skills improve over time.


PyPI version Python 3.12 License Docs

Documentation · Quick Start · PyPI · Examples · Issues


Start here

agloom is a Python framework for production-minded agents on LangChain / LangGraph. You describe the model and tools; agloom picks how to run the task (single-shot, ReAct, supervisor-style delegation, pipelines, and more), tracks steps and tokens, and can learn reusable skills from what worked.

If you already use LangChain’s agent APIs, think of create_agent as your main entrypoint — with orchestration, memory, streaming, and safety knobs in one place.

Install

pip install agloom
# optional extras, e.g. Groq:
pip install agloom[groq]

Your first agent

import asyncio
from langchain_groq import ChatGroq
from agloom import create_agent

async def main():
    llm = ChatGroq(model="meta-llama/llama-4-scout-17b-16e-instruct")
    agent = await create_agent(model=llm, name="my-agent")
    result = await agent.ainvoke("What causes auroras?")
    print(result.output)

asyncio.run(main())

create_agent is async (use await). From synchronous code, use create_agent_sync.

Next steps: Why agloom? · Patterns explained · All parameters


What you get (in plain language)

You want to… agloom helps by…
Ship faster Picking a strategy per query instead of hand-writing routers and graphs
Keep context Session memory by default; optional long-term memory and skills
Show progress Token streaming plus structured events for “thinking” / tool UIs
Stay safe Human-in-the-loop levels, timeouts, retries, rate limits — configurable
Improve over time Skill library and feedback hooks so behavior compounds

For the full feature tour, see What you get in the docs — the README stays short on purpose.


agloom CLI & web workspace

  • Terminal: the agloom CLI (npm agloom-cli, repo agloom_cli/) is the terminal client — React-based UI. From that folder: npm installnpm run buildnpm start. It talks to agloom-runtime over AGP (stdio by default). CLI quick start
  • Browser: agloom_web/ is the Vite workspace for sessions and observability — same idea, run commands inside that folder.

PyPI’s agloom package includes the library and agloom-runtime. The agloom command prints a short pointer to the agloom CLI (repo folder agloom_cli/) for backwards compatibility.


Learn more (documentation hub)

Guide What it’s for
Quick Start Smallest path to a running agent
Execution patterns How routing works (conceptual + diagrams)
Streaming & events Responsive UI patterns
Production Deploying, testing, operating
Errors & fixes When something goes wrong

Requirements

  • Python 3.12.x (see pyproject.toml on GitHub for the exact pin)
  • Node.js ≥ 24.15.0 — only if you hack on agloom_cli/ or agloom_web/
  • An LLM API key (Groq, OpenAI, NVIDIA, Hugging Face, or another LangChain-compatible provider)

Contributing & license

Contributions welcome — see CONTRIBUTING.md.

Licensed under Apache 2.0.


agloom

agloom is built by MEDHIRA

hello.medhira@gmail.com · GitHub · PyPI

Founded by S Muni Harish

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

agloom-0.1.73.tar.gz (427.5 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

agloom-0.1.73-py3-none-any.whl (338.8 kB view details)

Uploaded Python 3

File details

Details for the file agloom-0.1.73.tar.gz.

File metadata

  • Download URL: agloom-0.1.73.tar.gz
  • Upload date:
  • Size: 427.5 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: uv/0.11.6 {"installer":{"name":"uv","version":"0.11.6","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":null,"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":null}

File hashes

Hashes for agloom-0.1.73.tar.gz
Algorithm Hash digest
SHA256 adbb6afc03a059f2bb98127ff24c773999d6da6b11b5a7f9a83ac4b8bacd6d35
MD5 647b90d7ad3a9738aab72ad35adf7c08
BLAKE2b-256 1a505f8f4d44f8b55dc0597b0b0baf943e42eedd6c4b01ae1c68fd3f5fdcc17b

See more details on using hashes here.

File details

Details for the file agloom-0.1.73-py3-none-any.whl.

File metadata

  • Download URL: agloom-0.1.73-py3-none-any.whl
  • Upload date:
  • Size: 338.8 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: uv/0.11.6 {"installer":{"name":"uv","version":"0.11.6","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":null,"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":null}

File hashes

Hashes for agloom-0.1.73-py3-none-any.whl
Algorithm Hash digest
SHA256 7d8647457449c65fa4efb8bcc6258a78148786483cfa41fee5c60e6ac57a9e5e
MD5 7a3d558332f0266836ddd96d496d7fd9
BLAKE2b-256 18f3ccd4b7d085bbd3ef5c882078938be6528d144cf7886855f0097f9114d521

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page