Skip to main content

LLM Benchmark for Throughputs via Ollama

Project description

llm-benchmark (ollama-benchmark)

LLM Benchmark for Throughput via Ollama (Local LLMs)

Installation Steps

pip install llm-benchmark

Usage for general users directly

llm_benchmark run

Installation and Usage in Video format

llm-benchmark

It's tested on Python 3.9 and above.

ollama installation with the following models installed

7B model can be run on machines with 8GB of RAM

13B model can be run on machines with 16GB of RAM

Usage explaination

On Windows, Linux, and macOS, it will detect memory RAM size to first download required LLM models.

When memory RAM size is greater than or equal to 4GB, but less than 7GB, it will check if gemma:2b exist. The program implicitly pull the model.

ollama pull gemma:2b

When memory RAM size is greater than 7GB, but less than 15GB, it will check if these models exist. The program implicitly pull these models

ollama pull gemma:2b
ollama pull gemma:7b
ollama pull mistral:7b
ollama pull llama2:7b
ollama pull llava:7b

When memory RAM siz is greater than 15GB, it will check if these models exist. The program implicitly pull these models

ollama pull gemma:2b
ollama pull gemma:7b
ollama pull mistral:7b
ollama pull llama2:7b
ollama pull llama2:13b
ollama pull llava:7b
ollama pull llava:13b

Python Poetry manually(advanced) installation

https://python-poetry.org/docs/#installing-manually

For developers to develop new features on Windows Powershell or on Ubuntu Linux or macOS

python3 -m venv .venv
. ./.venv/bin/activate
pip install -U pip setuptools
pip install poetry

Usage in Python virtual environment

poetry shell
poetry install
llm_benchmark hello jason

Example #1 send systeminfo and benchmark results to a remote server

llm_benchmark run

Example #2 Do not send systeminfo and benchmark results to a remote server

llm_benchmark run --no-sendinfo

Example #3 Benchmark run on explicitly given the path to the ollama executable (When you built your own developer version of ollama)

llm_benchmark run --ollamabin=~/code/ollama/ollama

Reference

Ollama

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

llm_benchmark-0.3.13.tar.gz (2.1 MB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

llm_benchmark-0.3.13-py3-none-any.whl (2.1 MB view details)

Uploaded Python 3

File details

Details for the file llm_benchmark-0.3.13.tar.gz.

File metadata

  • Download URL: llm_benchmark-0.3.13.tar.gz
  • Upload date:
  • Size: 2.1 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.0.0 CPython/3.12.2

File hashes

Hashes for llm_benchmark-0.3.13.tar.gz
Algorithm Hash digest
SHA256 89f627d1ca4a1b8b899258ddf20fad8780c14f36bc34ff29ed358f1ae8dcabfd
MD5 6a560a2522ebed84d487669c036016b0
BLAKE2b-256 77e2513c26367daedd05345041fe01616cda0b5e8315945d2ac3fb3a50ddc8de

See more details on using hashes here.

File details

Details for the file llm_benchmark-0.3.13-py3-none-any.whl.

File metadata

  • Download URL: llm_benchmark-0.3.13-py3-none-any.whl
  • Upload date:
  • Size: 2.1 MB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.0.0 CPython/3.12.2

File hashes

Hashes for llm_benchmark-0.3.13-py3-none-any.whl
Algorithm Hash digest
SHA256 3d634ff1ff4ebd022f2968bc62328e730e760652f0a7469eda88870baf8a6610
MD5 4bff33e3ebea5958cbda0e4b41c3121c
BLAKE2b-256 ba16bd30fb0fcf1c1becdabee35f1be045b9d6d2cec8c6b03631279bba57098e

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page