Skip to main content

LLM-ify: create LLM-ready markdown and text from websites.

Project description

LLM-ify

LLM-ify

LLM-ify generates llms.txt and llms-full.txt plus full markdown/text captures for single pages or entire websites, so your LLM has clean, readable context.

What is llms.txt?

llms.txt is a standardized format for making website content more accessible to Large Language Models (LLMs). It provides:

  • llms.txt: A concise index of all pages with titles and descriptions
  • llms-full.txt: Complete content of all pages for comprehensive access

Features

  • Full website or single-page capture
  • Text + Markdown output
  • LLM-friendly index (llms.txt) + full corpus (llms-full.txt)
  • OpenAI or OpenRouter support
  • Interactive terminal UI

Prerequisites

  • Python 3.7+
  • OpenAI API key (Get one here)
  • Crawl4AI setup (see installation)

Quick Setup (Recommended)

Install from PyPI and launch:

pip install llmify-cli
llmify

Run the quickstart script for your OS:

# Windows (PowerShell)
scripts\quickstart.ps1

# macOS/Linux
./scripts/quickstart.sh

This creates a venv, installs deps, runs crawl4ai-setup, and launches the TUI.

Every time after this just activate your venv and run

python main.py

Manual Setup

  1. Clone the repository:
git clone https://github.com/Chillbruhhh/LLM-ify.git
cd LLM-ify
  1. Create and activate a virtual environment:
# Windows
python -m venv venv
venv\Scripts\activate

# macOS/Linux
python3 -m venv venv
source venv/bin/activate
  1. Install dependencies:
pip install -r requirements.txt
  1. Run the setup command:
crawl4ai-setup

This installs browser dependencies needed for Crawl4AI.

  1. Set up your OpenAI API key:

    Option A: Using .env file (recommended)

    cp .env.example .env
    # Edit .env and configure:
    # - Add OPENAI_API_KEY (required)
    

    Option B: Using environment variables

    export OPENAI_API_KEY="your-openai-api-key"
    

    Option C: Using command line arguments (See the TUI for input fields)

OpenRouter (Optional)

LLM-ify can also use OpenRouter by setting these environment variables:

OPENROUTER_API_KEY="your-openrouter-api-key"
OPENAI_BASE_URL="https://openrouter.ai/api/v1"
OPENAI_MODEL="openai/gpt-4.1-nano"

If OPENROUTER_API_KEY is set and OPENAI_API_KEY is empty, LLM-ify will use the OpenRouter key automatically.

Usage (TUI)

Launch the terminal UI:

python main.py

Enter a URL, choose a mode (full website or single page), and run. Settings are saved in config.json automatically.

Output Format

llms.txt

# https://example.com llms.txt

- [Page Title](https://example.com/page1): Brief description of the page content here
- [Another Page](https://example.com/page2): Another concise description of page content

llms-full.txt

# https://example.com llms-full.txt

<|llm-ify-page-1-lllmstxt|>
## Page Title
Full markdown content of the page...

<|llm-ify-page-2-lllmstxt|>
## Another Page
Full markdown content of another page...

Output Locations

Output files are written under collected-texts/crawl4ai-<domain>/ by default and split into txt-output and md-output folders. Example:

collected-texts/crawl4ai-docs.example.com/txt-output/llms.txt
collected-texts/crawl4ai-docs.example.com/txt-output/llms-full.txt
collected-texts/crawl4ai-docs.example.com/md-output/llms.md
collected-texts/crawl4ai-docs.example.com/md-output/llms-full.md
collected-texts/crawl4ai-docs.example.com/seeds.json

License

PolyForm Noncommercial - see LICENSE for details.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

llmify_cli-0.1.0.tar.gz (25.9 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

llmify_cli-0.1.0-py3-none-any.whl (24.8 kB view details)

Uploaded Python 3

File details

Details for the file llmify_cli-0.1.0.tar.gz.

File metadata

  • Download URL: llmify_cli-0.1.0.tar.gz
  • Upload date:
  • Size: 25.9 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.10

File hashes

Hashes for llmify_cli-0.1.0.tar.gz
Algorithm Hash digest
SHA256 1f37b623a30a1863ac845444acb6327bed329c73ed2dbebbc2c1c03bdedab901
MD5 7f95316288d6f0185799cc80ed7e3fd4
BLAKE2b-256 5fe1ea12014e5b582a8dd8203f9ed3cc80d340c58c2306d51d8251afa11d20c0

See more details on using hashes here.

File details

Details for the file llmify_cli-0.1.0-py3-none-any.whl.

File metadata

  • Download URL: llmify_cli-0.1.0-py3-none-any.whl
  • Upload date:
  • Size: 24.8 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.10

File hashes

Hashes for llmify_cli-0.1.0-py3-none-any.whl
Algorithm Hash digest
SHA256 967acd8dfb9a4372b598aa1ad9bf2885fa747d0fee7ff47927097ca341850161
MD5 c05f504acfb7e7942760caf7aca6bd46
BLAKE2b-256 985f2dff7fc279050038fd03fc012a1817cb482b7d919bd174ddf555ae13b93a

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page