GPU benchmarking tool for training and inference
Project description
GPU Benchmark by United Compute
A simple CLI tool to benchmark your GPU's performance across various AI models (Stable Diffusion, LLMs, etc.) and compare results in our global benchmark results.
Installation
pip install gpu-benchmark
Usage
Run the benchmark (takes 5 minutes after the pipeline is loaded):
gpu-benchmark
Available Benchmarks
You can specify which model to benchmark using the --model flag:
Stable Diffusion 1.5 (Default)
gpu-benchmark --model stable-diffusion-1-5
Qwen 3.0 6B (LLM Inference)
gpu-benchmark --model qwen3-0-6b
nanoGPT (LLM Training)
gpu-benchmark --model nanogpt-train
Optional Arguments
If you're running on a cloud provider, specify it with the --provider flag:
gpu-benchmark --provider runpod
For multi-GPU systems, you can select a specific GPU like this:
- Using the
--gpuflag:
gpu-benchmark --gpu 1 # Uses GPU index 1
The tool will:
- Load the selected model (Stable Diffusion, Qwen, or nanoGPT)
- Run the benchmark for 5 minutes
- Track performance metrics (throughput/iterations) and GPU temperature
- Upload results to the United Compute Benchmark Results
What it measures
- Benchmark Score: Number of iterations or images generated in 5 minutes (model-dependent)
- GPU Model: The specific model of your GPU (e.g., NVIDIA GeForce RTX 4090)
- Max Heat: Maximum GPU temperature reached (°C)
- Avg Heat: Average GPU temperature during the benchmark (°C)
- Country: Your location (detected automatically)
- GPU Power: Power consumption in watts (W)
- GPU Memory: Total GPU memory in gigabytes (GB)
- Platform: Operating system information
- Acceleration: CUDA version
- PyTorch Version: PyTorch library version
Requirements
- CUDA-compatible NVIDIA GPU
- Python 3.8+
Links
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file gpu_benchmark-0.4.3.tar.gz.
File metadata
- Download URL: gpu_benchmark-0.4.3.tar.gz
- Upload date:
- Size: 16.7 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.11.13
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
0610cb02d4d055354fdc4715baa0a324fdc76c15905f518cfd9f7617f9418722
|
|
| MD5 |
ae546b554b4c75558f31d3959845c384
|
|
| BLAKE2b-256 |
676de1fa2dabda3fd198463fe2e919d1a7506da7f199bcdb7ffb2bcb5109e16a
|
File details
Details for the file gpu_benchmark-0.4.3-py3-none-any.whl.
File metadata
- Download URL: gpu_benchmark-0.4.3-py3-none-any.whl
- Upload date:
- Size: 20.1 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.11.13
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
78bafe70a676aac09a451c08b48b2047a576712a26893d0d8a1c8a009d226a62
|
|
| MD5 |
3020dc53e761fa7c1324b51248eb2873
|
|
| BLAKE2b-256 |
97d6734ecb78b513476e325e21899641e903aa25c7ec601846d289c7b970e427
|