Skip to main content

Background AI assistant working as a Telegram bot, built specifically for document-related use cases

Project description

LlamaGram

LlamaGram is an AI agent inspired by OpenClaw (formerly known as MoltBot or ClawdBot), which focuses on document-related tasks.

It runs as a Telegram bot as comes with a CLI interface to set up the necessary environment variables.

Prerequisites

  • Python (if setting up the bot natively, preferably with uv) or Docker (if deploying with Docker)
  • A Telegram Bot Token, in order to connect to Telegram. Follow this guide on how to create your Telegram bot with BotFather.
  • A LlamaCloud API key, in order to give LlamaGram document-processing capabilities. Sign up on LlamaCloud here.
  • An API key for Google, OpenAI or Anthropic (you can choose one among the three or swap between different providers)

Installation

Install the bot natively:

# uv (recommended)
uv tool install llamagram 
# pip
pip install llamagram

Pull the docker image (only works for AMD64-compatible platforms):

docker pull ghcr.io/astrabert/llamagram:main

Setup

Through environment variables, you can customize the setup of LlamaGram:

  • LLAMAGRAM_LLM_PROVIDER: LLM provider (choose between google, anthropic and openai). Default is openai
  • LLAMAGRAM_LLM_MODEL: LLM model (choose among available models). Default is gpt-4.1

You then need to set three required env variables:

  • LLAMAGRAM_LLM_API_KEY: API key for the LLM (you can also use OPENAI_API_KEY, GOOGLE_API_KEY or ANTHROPIC_API_KEY, depending on the provider).
  • TELEGRAM_BOT_TOKEN: token for the Telegram bot
  • LLAMA_CLOUD_API_KEY: API key for LlamaCloud

You can use the setup wizard to configure LlamaGram interactively on the terminal:

llamagram setup --interactive

Or pass options from CLI:

llamagram setup --provider google \
    --model gemini-3-flash-preview \
    --api-key $GOOGLE_API_KEY \
    --llama-cloud-key $LLAMA_CLOUD_API_KEY \
    --telegram-token $TELEGRAM_BOT_TOKEN

This will create a .env file with the necessary variables, which will be loaded by LlamaGram at runtime (make sure not to share it with anyone).

Run

Run LlamaGram as a CLI app:

llamagram run 

You can set the --log-level option, if you wish to have more or less logging.

Run LlamaGram in a Docker container referencing a .env file:

docker run ghcr.io/astrabert/llamagram:main --env-file=".env"

Or, setting env varaibles directly (not recommended):

docker run ghcr.io/astrabert/llamagram:main \
    --env="LLAMAGRAM_LLM_PROVIDER=openai" \
    --env="LLAMAGRAM_LLM_MODEL=gpt-4.1"\
    --env="LLAMAGRAM_LLM_API_KEY=sk-xxx" \
    --env="LLAMA_CLOUD_API_KEY=llx-xxx" \
    --env="TELEGRAM_BOT_TOKEN=tok-xxx"

Use as a Telegram Bot

When on Telegram, you can perform two actions:

  • Sending PDF files, which will be downloaded by the bot
  • Sending text messages, which will work as prompts for the bot to start a new task

With /start command, you will have a welcome message explaining how to use the bot

How LlamaGram Works

LlamaGram is a generalist AI agent based on three main principles:

  • LlamaIndex Agent Workflows: a powerful workflow engine that allows event-driven, stepwise execution of specific tasks and functions. LlamaGram uses a cyclic workflow to go through thinking, tool-calling and observing repeatedly until it produces its final output.
  • Structured outputs: the LLM underlying the agent is forced to produce JSON outputs that comply with certain schemas (a tool call, a thought, an observation...): outputs are produced informed by the previous chat history, and based on context about available tools and specific tasks the agent has to perform.
  • Security by design: the agent does not have access to your real filesystem, but it does have access to a virtualized copy of it provided through AgentFS. PDFs sent over Telegram are also not downloaded into your real filesystem, but written within AgentFS. Files such as .envs or other popular credential files (.npmrc, .pypirc, .netrc) are excluded from the virtual filesystem, and thus unaccessible to the agent. The agent cannot use bash commands (it has access to filesystem-based tools like read/write/edit/grep/glob for AgentFS) to avoid it being able to perform destructive or vulnerable operations.

Here is what happens when you send a prompt to LlamaGram:

Flowchart LlamaGram

Along with the final response, the agent will also send you a report of everything it did during its session as a markdown file (namedd session-<random-id>-report.md).

License

This package is provided under MIT License

Contributing

For contributions, refer to the contributing guide

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

llamagram-0.1.0b0.tar.gz (9.5 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

llamagram-0.1.0b0-py3-none-any.whl (12.6 kB view details)

Uploaded Python 3

File details

Details for the file llamagram-0.1.0b0.tar.gz.

File metadata

  • Download URL: llamagram-0.1.0b0.tar.gz
  • Upload date:
  • Size: 9.5 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: uv/0.9.30 {"installer":{"name":"uv","version":"0.9.30","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Ubuntu","version":"24.04","id":"noble","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}

File hashes

Hashes for llamagram-0.1.0b0.tar.gz
Algorithm Hash digest
SHA256 b4ad429781c8632fcdee6fce56188046d2c5f34e2afe0aca7b016b80e2a771a5
MD5 f6f3e3f52ea4b825c83138a7fb011f8a
BLAKE2b-256 5d049e0a51be9533027885a682d006a37150cdc8a5a40f3a4b260b0331a8d877

See more details on using hashes here.

File details

Details for the file llamagram-0.1.0b0-py3-none-any.whl.

File metadata

  • Download URL: llamagram-0.1.0b0-py3-none-any.whl
  • Upload date:
  • Size: 12.6 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: uv/0.9.30 {"installer":{"name":"uv","version":"0.9.30","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Ubuntu","version":"24.04","id":"noble","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}

File hashes

Hashes for llamagram-0.1.0b0-py3-none-any.whl
Algorithm Hash digest
SHA256 207f588b85487da1c54b3ea0e9f0affdb38bcb3f16ee4106e3c3bf2d2add6d73
MD5 40b24b6b163f385c2a7f49768b0d25b3
BLAKE2b-256 5b6169a8faaf7670606bcaa4704854aa098bc6224f8cfd1c85d5afdcc03973e6

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page