Skip to main content

Background AI assistant working as a Telegram bot, built specifically for document-related use cases

Project description

LobsterX🦞

LobsterX is an AI agent inspired by OpenClaw (formerly known as MoltBot or ClawdBot), which focuses on document-related tasks.

It runs as a Telegram bot as comes with a CLI interface to set up the necessary environment variables.

Prerequisites

  • Python (if setting up the bot natively, preferably with uv) or Docker (if deploying with Docker)
  • A Telegram Bot Token, in order to connect to Telegram. Follow this guide on how to create your Telegram bot with BotFather.
  • A LlamaCloud API key, in order to give LobsterX document-processing capabilities. Sign up on LlamaCloud here.
  • An API key for Google, OpenAI or Anthropic (you can choose one among the three or swap between different providers)

Installation

Install the bot natively:

# uv (recommended)
uv tool install lobsterx --prerelease=allow
# pip
pip install lobsterx

Pull the docker image (only works for AMD64-compatible platforms):

docker pull ghcr.io/astrabert/lobsterx:main

Setup

Through environment variables, you can customize the setup of LobsterX:

  • LOBSTERX_LLM_PROVIDER: LLM provider (choose between google, anthropic and openai). Default is openai
  • LOBSTERX_LLM_MODEL: LLM model (choose among available models). Default is gpt-4.1

You then need to set three required env variables:

  • LOBSTERX_LLM_API_KEY: API key for the LLM (you can also use OPENAI_API_KEY, GOOGLE_API_KEY or ANTHROPIC_API_KEY, depending on the provider).
  • TELEGRAM_BOT_TOKEN: token for the Telegram bot
  • LLAMA_CLOUD_API_KEY: API key for LlamaCloud

If you wish to setup LobsterX as an API server, you will need to set an API key that only you can use to interact with it, set in the environment as LOBSTERX_SERVER_KEY. The key has to be at least 32 charachters long and contain only lowercase and uppercase alphanumeric characters, - and _.

You can use the setup wizard to configure LobsterX interactively on the terminal:

lobsterx setup --interactive

Or pass options from CLI:

lobsterx setup --provider google \
    --model gemini-3-flash-preview \
    --api-key $GOOGLE_API_KEY \
    --llama-cloud-key $LLAMA_CLOUD_API_KEY \
    --telegram-token $TELEGRAM_BOT_TOKEN \
    --server-key $SERVER_KEY

This will create a .env file with the necessary variables, which will be loaded by LobsterX at runtime (make sure not to share it with anyone).

If you wish to further customize the instructions that LobsterX has access to, you can use an AGENTS.md file, saved under the same directory where the agent process is running.

Run

As a Telegram Bot

Run LobsterX as a Telegram Bot:

lobsterx run 

You can set the --log-level option, if you wish to have more or less logging.

Run LobsterX in a Docker container referencing a .env file:

docker run ghcr.io/astrabert/lobsterx:main --env-file=".env"

Or, setting env varaibles directly (not recommended):

docker run ghcr.io/astrabert/lobsterx:main \
    --env="LOBSTERX_LLM_PROVIDER=openai" \
    --env="LOBSTERX_LLM_MODEL=gpt-4.1"\
    --env="LOBSTERX_LLM_API_KEY=sk-xxx" \
    --env="LLAMA_CLOUD_API_KEY=llx-xxx" \
    --env="TELEGRAM_BOT_TOKEN=tok-xxx"

When on Telegram, you can perform two actions:

  • Sending PDF files, which will be downloaded by the bot
  • Sending text messages, which will work as prompts for the bot to start a new task

With /start command, you will have a welcome message explaining how to use the bot

As an API server

To run as an API server, you need to specify a series of options that are necessary for authentication, rate limiting and CORS.

  • For authentication, you need to set the LOBSTERX_SERVER_KEY within the environment or in a .env file in the same working directory as the agent
  • For CORS, you can set a list of allowed origins
  • For rate limiting, you can set the maximum limits of file uploads, task creations, task polling and task deletion per minute

In addition to these, you will also need to provide the host (0.0.0.0 e.g.), port (8000 e.g.) and protocol (http or https) on which the server will run.

You can provide all of these details directly from the CLI:

lobsterx serve \
    --file-downloads-per-minute 300 \
    --create-tasks-per-minute 60 \
    --delete-tasks-per-minute 60 \
    --poll-tasks-per-minute 300 \
    --bind 0.0.0.0 \
    --port 8000 \
    --protocol http \
    --allow https://example.com \
    --allow https://anotherexample.com

All of these options have sensible defaults, but personalization is always recommended

Or create a JSON configuration (as in thie example) following this specification:

{
  "allow_origins": [],
  "file_downloads_per_minute": 300,
  "create_tasks_per_minute": 60,
  "delete_tasks_per_minute": 60,
  "poll_tasks_per_minute": 300,
  "host": "0.0.0.0",
  "port": 8000,
  "protocol": "http"
}

And provide it to the CLI:

lobsterx serve --config config.api.json

The configuration approach is recommended, as it can be re-use through different API-related commands.

Once you are serving your API through lobsterx serve, you can:

  • Upload files, by sending a POST request to /files
  • Create tasks, by sending a POST request to /task
  • Get the status of a task, by sending a GET request to /task/{task_id}
  • Cancel a task, by sending a DELETE request to /task/{task_id}

You don't have to do this through raw API calls, the LobsterX CLI provides several commands to perform these operations on your behalf:

# upload a file
lobsterx upload-file path/to/file.pdf --config config.api.json # pass the server configuration
# start a task
lobsterx create-task "Your prompt" --config config.api.json # this will return a task ID
# check the status of a task
lobsterx get-task some-task-id --config config.api.json
# cancel a task 
lobsterx cancel-task some-task-id --config config.api.json
# wait until a task is complete
lobsterx wait-task some-task-id --config config.api.json --polling-interval 2.0 --max-attempts 900 --verbose

How LobsterX Works

LobsterX is a generalist AI agent based on three main principles:

  • LlamaIndex Agent Workflows: a powerful workflow engine that allows event-driven, stepwise execution of specific tasks and functions. LobsterX uses a cyclic workflow to go through thinking, tool-calling and observing repeatedly until it produces its final output.
  • Structured outputs: the LLM underlying the agent is forced to produce JSON outputs that comply with certain schemas (a tool call, a thought, an observation...): outputs are produced informed by the previous chat history, and based on context about available tools and specific tasks the agent has to perform.
  • Security by design: the agent does not have access to your real filesystem, but it does have access to a virtualized copy of it provided through AgentFS. PDFs sent over Telegram are also not downloaded into your real filesystem, but written within AgentFS. Files such as .envs or other popular credential files (.npmrc, .pypirc, .netrc) are excluded from the virtual filesystem, and thus unaccessible to the agent. The agent cannot use bash commands (it has access to filesystem-based tools like read/write/edit/grep/glob for AgentFS) to avoid it being able to perform destructive or vulnerable operations.

Here is what happens when you send a prompt to LobsterX:

Flowchart LobsterX

Along with the final response, the agent will also send you a report of everything it did during its session as a markdown file (namedd session-<random-id>-report.md).

The API server

While sharing the core desing principles outlined above, the API server has some more features related to the data flow:

  • When a POST request to the /tasks endpoint (task creation) is made, a new asyncio.Task is spawned and stored within a in-memory task manager, using a locked dictionary to associate a task ID with an async Task.
  • When a GET request is sent to /task/{task_id}, the task manager provides details on the status of the task (success, failed, cancelled, pending). If the task was succesfull, failed or was cancelled, it is removed from the dictionary.
  • When a DELETE request is sent to /task/{task_id}, the async Task is cancelled and removed from the dictionary

Besides the Task Manager, the API server uses an in-memory rate limiter (fastapi-throttle) and Starlette CORS and Auth middleawares to provide authentication (through a Bearer token provided with an Authorization header) and CORS servicres.

License

This package is provided under MIT License

Contributing

For contributions, refer to the contributing guide

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

lobsterx-0.2.1b0.tar.gz (16.5 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

lobsterx-0.2.1b0-py3-none-any.whl (21.8 kB view details)

Uploaded Python 3

File details

Details for the file lobsterx-0.2.1b0.tar.gz.

File metadata

  • Download URL: lobsterx-0.2.1b0.tar.gz
  • Upload date:
  • Size: 16.5 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: uv/0.10.8 {"installer":{"name":"uv","version":"0.10.8","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Ubuntu","version":"24.04","id":"noble","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}

File hashes

Hashes for lobsterx-0.2.1b0.tar.gz
Algorithm Hash digest
SHA256 47e31d0948d31a410ed2446287eef026d6d3d2413fac3a395e3111c9ef71a013
MD5 7ce367df35f31568c6796a586f415d37
BLAKE2b-256 8b0ffffb1f102f7d8fbdba1bf19d150862d0e2aeb42b6afaf2cfc67247d43379

See more details on using hashes here.

File details

Details for the file lobsterx-0.2.1b0-py3-none-any.whl.

File metadata

  • Download URL: lobsterx-0.2.1b0-py3-none-any.whl
  • Upload date:
  • Size: 21.8 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: uv/0.10.8 {"installer":{"name":"uv","version":"0.10.8","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Ubuntu","version":"24.04","id":"noble","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":true}

File hashes

Hashes for lobsterx-0.2.1b0-py3-none-any.whl
Algorithm Hash digest
SHA256 bddd897ff4f88a02f30f7d8283954b1715e706d8612133036479de8434d5b3df
MD5 886b7f55cd4173a1d78e3f4972e9aac3
BLAKE2b-256 49604b263cf7285a8cbf77aa68f14207d137c854cf1087a5c61879e03223d789

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page