A framework for holistic evaluation of LLM Inference Systems
Project description
Veeksha
Veeksha is a high-fidelity benchmarking framework for LLM inference systems. Whether you're optimizing a production deployment, comparing serving backends, or running capacity planning experiments, Veeksha lets you measure what matters to you: realistic multi-turn conversations, agentic workflows, high-frequency stress tests, or targeted microbenchmarks. One tool, any workload.
From isolated requests to complex agentic sessions, Veeksha captures the full complexity of modern LLM workloads.
👉 Why Veeksha? — Learn what sets Veeksha apart
📚 Documentation — Full guides and API reference
Quick start
In a fresh environment (Python 3.14t recommended for true parallelism):
Install from PyPI:
pip install veeksha
Run a benchmark against an OpenAI-compatible endpoint:
python -Xgil=0 -m veeksha.benchmark \
--client-type openai_chat_completions \
--openai-chat-completions-client-api-base http://localhost:8000/v1 \
--openai-chat-completions-client-model meta-llama/Llama-3.2-1B-Instruct \
--traffic-scheduler-type rate \
--rate-traffic-scheduler-interval-generator-type poisson \
--rate-traffic-scheduler-poisson-interval-generator-arrival-rate 5.0 \
--runtime-benchmark-timeout 60
Or use a YAML configuration file:
python -Xgil=0 -m veeksha.benchmark --benchmark-config-from-file my_benchmark.veeksha.yml
Installation from source
git clone https://github.com/project-vajra/veeksha.git
cd veeksha
# Install uv if needed
curl -LsSf https://astral.sh/uv/install.sh | sh
# Create environment (Python 3.14t recommended for true parallelism)
uv venv --python 3.14t
source .venv/bin/activate
uv pip install -e .
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file veeksha-0.2.0.tar.gz.
File metadata
- Download URL: veeksha-0.2.0.tar.gz
- Upload date:
- Size: 567.0 kB
- Tags: Source
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
bec242a2f8aa4eba9f2635e37ce1c767434f0e6126e70f0541c997319d8100e8
|
|
| MD5 |
7f1078aa1f314b5aef04407d813f2bb6
|
|
| BLAKE2b-256 |
b0f380878b396ec0c2de94e9190e292a385ecb2a4fb76d824fa0faed0b4cb394
|
Provenance
The following attestation bundles were made for veeksha-0.2.0.tar.gz:
Publisher:
publish_release.yml on project-vajra/veeksha
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
veeksha-0.2.0.tar.gz -
Subject digest:
bec242a2f8aa4eba9f2635e37ce1c767434f0e6126e70f0541c997319d8100e8 - Sigstore transparency entry: 837624207
- Sigstore integration time:
-
Permalink:
project-vajra/veeksha@4af8b6e81c1bdc3a24c2f9bd2faf17f87dba44c7 -
Branch / Tag:
refs/tags/v0.2.0 - Owner: https://github.com/project-vajra
-
Access:
private
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
publish_release.yml@4af8b6e81c1bdc3a24c2f9bd2faf17f87dba44c7 -
Trigger Event:
release
-
Statement type:
File details
Details for the file veeksha-0.2.0-py3-none-any.whl.
File metadata
- Download URL: veeksha-0.2.0-py3-none-any.whl
- Upload date:
- Size: 235.6 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
21223326480fdd110d90f8828241bca7aad52ce6eccbd076c18bd71291cd250b
|
|
| MD5 |
683c1d1acdf3b380068db873209b32ee
|
|
| BLAKE2b-256 |
238307c3fb9dd69d5dff7a053453c3324943546ef1ffb1560d09c3cf7288031b
|
Provenance
The following attestation bundles were made for veeksha-0.2.0-py3-none-any.whl:
Publisher:
publish_release.yml on project-vajra/veeksha
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
veeksha-0.2.0-py3-none-any.whl -
Subject digest:
21223326480fdd110d90f8828241bca7aad52ce6eccbd076c18bd71291cd250b - Sigstore transparency entry: 837624257
- Sigstore integration time:
-
Permalink:
project-vajra/veeksha@4af8b6e81c1bdc3a24c2f9bd2faf17f87dba44c7 -
Branch / Tag:
refs/tags/v0.2.0 - Owner: https://github.com/project-vajra
-
Access:
private
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
publish_release.yml@4af8b6e81c1bdc3a24c2f9bd2faf17f87dba44c7 -
Trigger Event:
release
-
Statement type: