Skip to main content

FlashInfer-Bench - AI for AI Infrastructure for Accelerating AI Deployment

Project description

FlashInfer-Bench logo

Documentation License PyPI

Building the Virtuous Cycle for AI-driven LLM Systems

Get Started | Documentation | Blogpost | Slack (#flashinfer-bench)

FlashInfer-Bench is a benchmark suite and production workflow designed to build a virtuous cycle of self-improving AI systems.

It is part of a broader initiative to build the virtuous cycle of AI improving AI systems — enabling AI agents and engineers to collaboratively optimize the very kernels that power large language models.

Installation

Install FlashInfer-Bench with pip:

pip install flashinfer-bench

Import FlashInfer-Bench:

import flashinfer_bench as fib

print(fib.__version__)

Get Started

This guide shows you how to use FlashInfer-Bench python module with the FlashInfer-Trace dataset.

FlashInfer Trace Dataset

We provide an official dataset called FlashInfer-Trace with kernels and workloads in real-world AI system deployment environments. FlashInfer-Bench can use this dataset to measure and compare the performance of kernels. It follows the FlashInfer Trace Schema.

The official dataset is on HuggingFace: https://huggingface.co/datasets/flashinfer-ai/flashinfer-trace

Collaborators

Our collaborators include:

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

flashinfer_bench-0.1.2.tar.gz (1.1 MB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

flashinfer_bench-0.1.2-py3-none-any.whl (145.9 kB view details)

Uploaded Python 3

File details

Details for the file flashinfer_bench-0.1.2.tar.gz.

File metadata

  • Download URL: flashinfer_bench-0.1.2.tar.gz
  • Upload date:
  • Size: 1.1 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/6.1.0 CPython/3.13.7

File hashes

Hashes for flashinfer_bench-0.1.2.tar.gz
Algorithm Hash digest
SHA256 3056ba7a9da4ac3cf6727c50038b9c7422c75f0c38e14e1e9a9aaa59a775d238
MD5 ede94d8f117f3c4aa0fb1a23d4922702
BLAKE2b-256 e5132f780c9366602992b034bd7d38fba896ada836f2c9d93f5025e4f523e938

See more details on using hashes here.

Provenance

The following attestation bundles were made for flashinfer_bench-0.1.2.tar.gz:

Publisher: build-and-upload-pypi.yml on flashinfer-ai/flashinfer-bench

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

File details

Details for the file flashinfer_bench-0.1.2-py3-none-any.whl.

File metadata

File hashes

Hashes for flashinfer_bench-0.1.2-py3-none-any.whl
Algorithm Hash digest
SHA256 4814cce18dcdb5db89c96d3411ae401e6b30ad8e5b727926eb902bc53f3ddb48
MD5 88edaf96ca4c19f39095109f0a413fcb
BLAKE2b-256 cddd0e47f64544943f17742506acd54a464e5e091ff36b049b9436d853deb871

See more details on using hashes here.

Provenance

The following attestation bundles were made for flashinfer_bench-0.1.2-py3-none-any.whl:

Publisher: build-and-upload-pypi.yml on flashinfer-ai/flashinfer-bench

Attestations: Values shown here reflect the state when the release was signed and may no longer be current.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page