Skip to main content

A unified library for creating, representing, and storing speculative decoding algorithms for LLM serving such as in vLLM.

Project description

Speculators logo

License Python Versions docs PyPI tests

Overview

Speculators is a library for training speculative decoding draft models that deploy directly to LLM inference engines like vLLM. Speculative decoding is a lossless technique that speeds up LLM inference by using a smaller, faster draft model (i.e. "the speculator") to propose tokens, which are then verified by the larger base model, reducing latency without compromising output quality. The speculator intelligently drafts multiple tokens ahead of time, and the base model verifies them in a single forward pass. This approach boosts performance without sacrificing output quality, as every accepted token is guaranteed to match what the main model would have generated on its own.

Speculators standardizes this process by providing a productionized end-to-end framework to train draft models with reusable formats and tools. Trained models can seamlessly run in vLLM, enabling the deployment of speculative decoding in production-grade inference servers.

Speculators user flow diagram


💬 Join us on the vLLM Community Slack and share your questions, thoughts, or ideas in:

  • #speculators
  • #feat-spec-decode

🎥 Watch our Office Hours presentation: Video | Slides


🚀 What's New!

Big updates have landed in Speculators! To get a more in-depth look, check out the Speculators documentation.

Some of the exciting new features include:

  • Gemma 4 Speculators: The RedHat team published speculators for Gemma 4 31B-it, including both DFlash and EAGLE-3 checkpoints, enabling production-grade speculative decoding for Gemma 4 models.
  • DFlash Training Algorithm: Added support for the DFlash training algorithm with anchored-block drafting, using auxiliary hidden states from multiple verifier layers. Includes CLI options for block size and max anchors, plus DFlash metrics, utilities, and draft model. DFlash models trained through Speculators can now run seamlessly in vLLM as of vLLM PR #38300.
  • Online Training Support: Added support for online training using the new vLLM hidden extraction system, enabling real-time hidden state generation during training without requiring separate offline data generation steps.

Key Features

  • Offline Training Data Generation using vLLM: Enable the generation of hidden states using vLLM. Data samples are saved to disk and can be used for draft model training.
  • Draft Model Training Support: E2E training support of single and multi-layer draft models. Training is supported for MoE, non-MoE, and Vision Language models.
  • Standardized, Extensible Format: Provides a Hugging Face-compatible format for defining speculative models, with tools to convert from external research repositories into a standard speculators format for easy adoption.
  • Seamless vLLM Integration: Built for direct deployment into vLLM, enabling low-latency, production-grade inference with minimal overhead.

[!TIP] Read more about Speculators features in this vLLM blog post.

Supported Models

The following table summarizes the models that have been trained end-to-end by our team as well as others in the roadmap:

Verifier Architecture Verifier Size Training Support vLLM Deployment Support
Llama 8B-Instruct EAGLE-3
70B-Instruct EAGLE-3
Qwen3 8B EAGLE-3
14B EAGLE-3
32B EAGLE-3
gpt-oss 20b EAGLE-3
120b EAGLE-3
Qwen3 MoE 30B-Instruct EAGLE-3
235B-Instruct EAGLE-3
235B EAGLE-3
Qwen3-VL 235B-A22B EAGLE-3
Mistral 3 Large 675B-Instruct EAGLE-3 ⏳
Gemma 4 31B-it EAGLE-3
DFlash
Gemma 4 MoE 26B-A4B-it EAGLE-3

✅ = Supported, ⏳ = In Progress, ❌ = Not Yet Supported

vLLM Inference

Models trained through Speculators can run seamlessly in vLLM using a simple vllm serve <speculator_model> command. This will run the model in vLLM using default arguments, defined in the speculator_config of the model's config.json.

vllm serve RedHatAI/Qwen3-8B-speculator.eagle3

Served models can then be benchmarked using GuideLLM. Below, we show sample benchmark results where we compare our speculator with its dense counterpart. We also additionally compare quantization to explore additional performance improvements by swapping the dense verifier, Qwen/Qwen3-8B with the quantized FP8 model, RedHatAI/Qwen3-8B-FP8-dynamic in the speculator_config.

GuideLLM Logo

Additional Utility Scripts

Getting Started

Installation

Prerequisites

Before installing, ensure you have the following:

  • Operating System: Linux or macOS
  • Python: 3.10 or higher
  • Package Manager: pip (recommended) or conda

Install from PyPI (Recommended)

Install the latest stable release from PyPI:

pip install speculators

Install from Source

For the latest development version or to contribute to the project:

git clone https://github.com/vllm-project/speculators.git
cd speculators

pip install -e .

For development with additional tools:

pip install -e ".[dev]"

Verify Installation

You can verify your installation by checking the version:

speculators --version

Or by importing the package in Python:

import speculators
print(speculators.__version__)

License

Speculators is licensed under the Apache License 2.0.

Cite

If you find Speculators helpful in your research or projects, please consider citing it:

@misc{speculators2025,
  title={Speculators: A Unified Library for Speculative Decoding Algorithms in LLM Serving},
  author={Red Hat},
  year={2025},
  howpublished={\url{https://github.com/vllm-project/speculators}},
}

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

speculators-0.5.0.tar.gz (101.1 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

speculators-0.5.0-py3-none-any.whl (112.4 kB view details)

Uploaded Python 3

File details

Details for the file speculators-0.5.0.tar.gz.

File metadata

  • Download URL: speculators-0.5.0.tar.gz
  • Upload date:
  • Size: 101.1 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for speculators-0.5.0.tar.gz
Algorithm Hash digest
SHA256 db6d5027494554dbf6819212578c43a8cc51971b24b9ee303d389f61da90f5ab
MD5 747b042b9af3560f0038c9266718aeef
BLAKE2b-256 a7dc064f8252d72ac03cae4aaa58c1af612e60848e24244492465c85e2a954d1

See more details on using hashes here.

Provenance

The following attestation bundles were made for speculators-0.5.0.tar.gz:

Publisher: speculators-upload.yml on neuralmagic/llm-compressor-testing

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file speculators-0.5.0-py3-none-any.whl.

File metadata

  • Download URL: speculators-0.5.0-py3-none-any.whl
  • Upload date:
  • Size: 112.4 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for speculators-0.5.0-py3-none-any.whl
Algorithm Hash digest
SHA256 24c1876ca7448b12b8d3be4c31cd736584d2248795024dd98e61570d02215c1d
MD5 5f308d633e4da915b100a503c754ac85
BLAKE2b-256 b12b3d83e581cbbc1ca47e7bfc88c7e79f5dbd5e14ae4404edf64646b6873cea

See more details on using hashes here.

Provenance

The following attestation bundles were made for speculators-0.5.0-py3-none-any.whl:

Publisher: speculators-upload.yml on neuralmagic/llm-compressor-testing

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page