Skip to main content

RapidFire AI: Rapid experimentation for easier, faster, and more impactful AI customization. Built for agentic RAG, context engineering, fine-tuning, and post-training of LLMs and other DL models.

Project description

RapidFire AI

Rapid AI Customization from RAG to Fine-Tuning

20x experimentation throughput of LLM pipelines faster, more systematic.

Try RAG on Colab    Try Fine-Tuning on Colab

PyPI version

RapidFire AI

Rapid experimentation for easier, faster, and more impactful AI customization. Built for agentic RAG, context engineering, fine-tuning, and post-training of LLMs and other DL models. Delivers 16-24x higher throughput without extra resources.

Overview

RapidFire AI is a new experiment execution framework that transforms your AI customization experimentation from slow, sequential processes into rapid, intelligent workflows with hyperparallelized execution, dynamic real-time experiment control, and automatic system optimization.

Usage workflow of RapidFire AI

RapidFire AI's adaptive execution engine allows interruptible, shard-based scheduling so you can compare many configurations concurrently, even on a single GPU (for self-hosted models) or a CPU-only machine (for closed model APIs) with dynamic real-time control over runs.

  • Hyperparallelized Execution: Higher throughput, simultaneous, data shard-at-a-time execution to show side-by-side differences.
  • Interactive Control (IC Ops): Stop, Resume, Clone-Modify, and optionally warm start runs in real-time from the dashboard.
  • Automatic Optimization: Intelligent single and multi-GPU orchestration to optimize utilization with minimal overhead for self-hosted models; intelligent token spend and rate limit apportioning for closed model APIs.

Shard-based concurrent execution (1 GPU)

For additional context, see the overview: RapidFire AI Overview

Getting Started

Prerequisites

Install and Get Started

# Ensure that python3 resolves to python3.12 if needed
python3 --version  # must be 3.12.x

python3 -m venv .venv
source .venv/bin/activate

pip install rapidfireai

rapidfireai --version
# Verify it prints the following:
# RapidFire AI 0.14.0

# Replace YOUR_TOKEN with your actual Hugging Face token
# https://huggingface.co/docs/hub/en/security-tokens
hf auth login --token YOUR_TOKEN

# Due to current issue: https://github.com/huggingface/xet-core/issues/527
pip uninstall -y hf-xet

# Depending on whether you want Fine-tuning/Post-Training or RAG/context eng., pick one of the remaining series of commands


# For Fine-tuning/Post-Training: Install specific dependencies and initialize rapidfireai

rapidfireai init
rapidfireai start

# It should print about 50 lines, including the following:
# ...
# RapidFire Frontend is ready
# Open your browser and navigate to: http://0.0.0.0:8853
# ...
# Press Ctrl+C to stop all services

# Forward this port if you installed rapidfireai on a remote machine
ssh -L 8853:localhost:8853 username@remote-machine

# Open an example notebook from ./tutorial_notebooks/[fine-tuning | post-training] and start experiment


# [OR]


# For RAG/Context Engineering Evals: Install specific dependencies and initialize rapidfireai
rapidfireai init --evals
rapidfireai start

# It should print about 50 lines, including the following:
# ...
# RapidFire Frontend is ready
# Open your browser and navigate to: http://0.0.0.0:8853
# ...
# Press Ctrl+C to stop all services

# For the RAG/context eng. notebooks, only jupyter is supported for now and must be started as follows
rapidfireai jupyter

# Forward these ports if you installed rapidfireai on a remote machine
ssh -L 8850:localhost:8850 -L 8851:localhost:8851 -L 8853:localhost:8853 username@remote-machine

# Open the URL provided by the jupyter notebook command above via your browser
# Open an example notebook from ./tutorial_notebooks/rag-contexteng/ and start experiment

Troubleshooting

For a quick system diagnostics report (Python env, relevant packages, GPU/CUDA, and key environment variables), run:

rapidfireai doctor

If you encounter port conflicts, you can kill existing processes:

lsof -t -i:8850 | xargs kill -9  # jupyter server
lsof -t -i:8851 | xargs kill -9  # dispatcher
lsof -t -i:8852 | xargs kill -9  # mlflow
lsof -t -i:8853 | xargs kill -9  # frontend server
lsof -t -i:8855 | xargs kill -9  # ray dashboard

Documentation

Browse or reference the full documentation, example use case tutorials, all API details, dashboard details, and more in the RapidFire AI Documentation.

Key Features

MLflow Integration

Full MLflow support for experiment tracking and metrics visualization. A named RapidFire AI experiment corresponds to an MLflow experiment for comprehensive governance

Interactive Control Operations (IC Ops)

First-of-its-kind dynamic real-time control over runs in flight. Can be invoked through the dashboard:

  • Stop active runs; puts them in a dormant state
  • Resume stopped runs; makes them active again
  • Clone and modify existing runs, with or without warm starting from parent's weights
  • Delete unwanted or failed runs

Multi-GPU Support

The Scheduler automatically handles multiple GPUs on the machine and divides resources across all running configs for optimal resource utilization.

Search and AutoML Support

Built-in procedures for searching over configuration knob combinations, including Grid Search and Random Search. Easy to integrate with AutoML procedures. Native support for some popular AutoML procedures and customized automation of IC Ops coming soon.

Directory Structure

rapidfireai/
├── automl/              # Search and AutoML algorithms for knob tuning
├── cli.py               # CLI script
├── evals
    ├── actors/          # Ray-based workers for doc and query processing  
    ├── data/            # Data sharding and handling
    ├── db/              # Database interface and SQLite operations
    ├── dispatcher/      # Flask-based web API for UI communication
    ├── metrics/         # Online aggregation logic and metrics handling
    ├── rag/             # Stages of RAG pipeline
    ├── scheduling/      # Fair scheduler for multi-config resource sharing
    └── utils/           # Utility functions and helper modules
├── experiment.py        # Main experiment lifecycle management
├── fit
    ├── backend/         # Core backend components (controller, scheduler, worker)
    ├── db/              # Database interface and SQLite operations
    ├── dispatcher/      # Flask-based web API for UI communication
    ├── frontend/        # Frontend components (dashboard, IC Ops implementation)
    ├── ml/              # ML training utilities and trainer classes
    └── utils/           # Utility functions and helper modules
└── utils.py             # Utility functions and helper modules

Architecture

RapidFire AI adopts a microservices-inspired loosely coupled distributed architecture with:

  • Dispatcher: Web API layer for UI communication
  • Database: SQLite for state persistence
  • Controller: Central orchestrator running in user process
  • Workers: GPU-based training processes (for SFT/RFT) or Ray-based Actors for doc and query processing (for RAG/context engineering)
  • Dashboard: Experiment tracking and visualization dashboard

This design enables efficient resource utilization while providing a seamless user experience for AI experimentation.

Components

Dispatcher

The dispatcher provides a REST API interface for the web UI. It can be run via Flask as a single app or via Gunicorn to have it load balanced. Handles interactive control features and displays the current state of the runs in the experiment.

Database

Uses SQLite for persistent storage of metadata of experiments, runs, and artifacts. The Controller also uses it to talk with Workers on scheduling state. A clean asynchronous interface for all DB operations, including experiment lifecycle management and run tracking.

Controller

Runs as part of the user’s console or Notebook process. Orchestrates the entire training lifecycle including model creation, worker management, and scheduling, as well as the entire RAG/context engineering pipeline for evals. The run_fit logic handles sample preprocessing, model creation for given knob configurations, worker initialization, and continuous monitoring of training progress across distributed workers. The run_evals logic handles data chunking, embedding, retrieval, reranking, context construction, and generation for inference evals.

Worker

Handles the actual model training and inference on the GPUs for run_fit and the data preprocessing and RAG inference evals for run_evals. Workers poll the Database for tasks, load dataset shards, and execute config-specific tasks: training runs with checkpointing (for SFT/RFT) and doc processing followed by query processing with online aggregation (for RAG/context eng. evals). Both also handle progress reporting. Currently expects any given model for given batch size to fit on a single GPU (for self-hosted models). Likewise, currently expects OpenAI API key provided to have sufficient balance for given evals workload.

Experiment

Manages the complete experiment lifecycle, including creation, naming conventions, and cleanup. Experiments are automatically named with unique suffixes if conflicts exist, and all experiment metadata is tracked in the Database. An experiment's running tasks are automatically cancelled when the process ends abruptly.

Dashboard

A fork of MLflow that enables full tracking and visualization of all experiments and runs for run_fit. It features a new panel for Interactive Control Ops that can be performed on any active runs. For run_evals the metrics are displayed in an auot-updated table on the notebook itself, while IC Ops panel also appears on the notebook itself.

Developing with RapidFire AI

Development prerequisites

TODO: This section needs updating

  • Python 3.12.x
  • Git
  • Ubuntu/Debian system (for apt package manager)
# Run these commands one after the other on a fresh Ubuntu machine

# install dependencies
sudo apt update -y

# clone the repository
git clone https://github.com/RapidFireAI/rapidfireai.git

# navigate to the repository
cd ./rapidfireai

# install basic dependencies
sudo apt install -y python3.12-venv
python3 -m venv .venv
source .venv/bin/activate
pip3 install ipykernel
pip3 install jupyter
pip3 install "huggingface-hub[cli]"
export PATH="$HOME/.local/bin:$PATH"
hf auth login --token <your_token>

# Due to current issue: https://github.com/huggingface/xet-core/issues/527
pip uninstall -y hf-xet

# checkout the main branch
git checkout main

# install the repository as a python package
pip3 install -r requirements.txt

# install node
curl -fsSL https://deb.nodesource.com/setup_22.x | sudo -E bash - && sudo apt-get install -y nodejs

# Install correct version of vllm and flash-attn
# uv pip install vllm=0.10.1.1 --torch-backend=cu126 or cu118
# uv pip install flash-attn==1.0.9 --no-build-isoloation or 2.8.3

# if running into node versioning errors, remove the previous version of node then run the lines above again
sudo apt-get remove --purge nodejs libnode-dev libnode72 npm
sudo apt autoremove --purge

# check installations
node -v # 22.x

# still inside venv, run the start script to begin all 3 servers
chmod +x ./rapidfireai/start_dev.sh
./rapidfireai/start_dev.sh start

# run the notebook from within your IDE
# make sure the notebook is running in the .venv virtual environment
# head to settings in Cursor/VSCode and search for venv and add the path - $HOME/rapidfireai/.venv
# we cannot run a Jupyter notebook directly since there are restrictions on Jupyter being able to create child processes

# VSCode can port-forward localhost:8853 where the rf-frontend server will be running

# for port clash issues -
lsof -t -i:8850 | xargs kill -9 # jupyter server
lsof -t -i:8851 | xargs kill -9 # dispatcher
lsof -t -i:8852 | xargs kill -9 # mlflow
lsof -t -i:8853 | xargs kill -9 # frontend
lsof -t -i:8855 | xargs kill -9 # ray console

RapidFireAI Environment Variables

RapidFire AI has sane defaults for most installations, if customization is needed the following operating system variables can be used to overwrite the defaults.

  • RF_HOME - Base RapidFire AI home directory (default: ${HOME}/rapidfireai on Non-Google Colab and /content/rapidfireai on Google Colab)
  • RF_LOG_PATH - Base directory to store log files (default: ${RF_HOME}/logs)
  • RF_EXPERIMENT_PATH - Base directory to store experiment work files (default: ${RF_HOME}/rapidfire_experiments)
  • RF_TENSORBOARD_LOG_DIR - Base directory for Tensorboard logs (default: ${RF_EXPERIMENT_PATH}/tensorboard_logs))
  • RF_LOG_FILENAME - Default log file name (default: rapidfire.log)
  • RF_TRAINING_LOG_FILENAME - Default training log file name (default: training.log)
  • RF_DB_PATH - Base directory for database files (default: ${RF_HOME}/db)
  • RF_MLFLOW_ENABLED - Enable MLFlow tracking backend
  • RF_TENSORBOARD_ENABLED - Enable Tensorboard tracking backend
  • RF_TRACKIO_ENABLED - Enable Trackio tracking backend
  • RF_COLAB_MODE - Whether running on colab (default: false on Non-Google Colab and true on Google Colab)
  • RF_TUTORIAL_PATH - Location that rapidfireai init copies tutorial_notebooks to (default: ./tutorial_notebooks)
  • RF_TEST_PATH - Location that rapidfireai --test-noteobooks copies test notebooks to (default: ./tutorial_notebooks/tests)
  • RF_JUPYTER_HOST - Host that rapidfireai jupyter creates a Jupyter listener for (default: 127.0.0.1)
  • RF_JUPYTER_PORT - Port that rapidfireai jupyter creates a Jupyter listener for (default: 8850)
  • RF_API_HOST - Host that rapidfireai start or Experiment creates an API listener for (default: 127.0.0.1)
  • RF_API_PORT - Port that rapidfireai start or Experiment creates an API listener for (default: 8851)
  • RF_MLFLOW_HOST - Host that rapidfireai start creates a MLFow listener for (default: 127.0.0.1)
  • RF_MLFLOW_PORT - Port that rapidfireai start creates a MLFow listener for (default: 127.0.0.1)
  • RF_FRONTEND_HOST - Host that rapidfireai start creates a Frontend listener for (default: 0.0.0.0)
  • RF_FRONTEND_PORT - Port that rapidfireai start creates a Frontend listener for (default: 0.0.0.0)
  • RF_RAY_HOST - Host that Experiment creates a Ray dashboard listener for (default: 127.0.0.1)
  • RF_RAY_PORT - Port that Experiment creates a Ray dashboard listener for (default: 127.0.0.1)
  • RF_TIMEOUT_TIME - Time in seconds that services wait to start (default: 30)
  • RF_PID_FILE - File to store process ids of started services (default: ${RF_HOME}/rapidfire_pids.txt)
  • RF_PYTHON_EXECUTABLE - Python executable (default: python3 falls back to python if not found)
  • RF_PIP_EXECUTABLE - pip executable (default: pip3 falls back to pip if not found)

Community & Governance

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

rapidfireai-0.14.0.tar.gz (46.9 MB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

rapidfireai-0.14.0-py3-none-any.whl (47.4 MB view details)

Uploaded Python 3

File details

Details for the file rapidfireai-0.14.0.tar.gz.

File metadata

  • Download URL: rapidfireai-0.14.0.tar.gz
  • Upload date:
  • Size: 46.9 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.12

File hashes

Hashes for rapidfireai-0.14.0.tar.gz
Algorithm Hash digest
SHA256 cdbe62ca283edc7aafd1272ed711f9eb11cd1b823b07cc184dd6df7c518c89e0
MD5 f012d5b89dce8e8a81f633fad1ce9aa5
BLAKE2b-256 e9f7319821f55575faf54f711fc1ef76b9e375e8b4e34f8895e5b46f416eda55

See more details on using hashes here.

File details

Details for the file rapidfireai-0.14.0-py3-none-any.whl.

File metadata

  • Download URL: rapidfireai-0.14.0-py3-none-any.whl
  • Upload date:
  • Size: 47.4 MB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.12.12

File hashes

Hashes for rapidfireai-0.14.0-py3-none-any.whl
Algorithm Hash digest
SHA256 c87a5bd81ec40b2c4b8b7f4708b13e71f0a82cfc9896d884d1325c34ad94693c
MD5 ff0c09f735bf326684f61ea9e8cbd1b1
BLAKE2b-256 dcd60c93a736fc291c75cac3999deb557ff384946e55709785b2fa541de11d71

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page