Client app for kalavai platform
Project description
⭐⭐⭐ Kalavai is offering GPU and other resources to interesting Open Source projects that want to push the envelope but are struggling to fund computing resources. If you are a research lab wanting to do research on LLMs, or a small startup trying to beat the tech giants with frugal models, we want to help Apply here
Kalavai: a platform to self-host AI on easy mode
AI in the cloud is not aligned with you, it's aligned with the company that owns it. Make sure you own your AI
Taming the adoption of self-hosted GenAI
Kalavai is an open source tool that turns any devices into a self-hosted AI platform. It aggregates resources from multiple machines, including cloud, on prem and personal computers, and is compatible with most model engines to make model deployment and orchestration simple and reliable.
What can Kalavai do?
Kalavai's goal is to make using self-hosted AI (GenAI models and agents) in real applications accessible and affordable to all.
Core features
- Manage multiple devices resources as one, wherever they come from (hybrid cloud, on prem, personal devices)
- Deploy open source models seamlessly across devices, with zero-cost migration
- Beyond LLMs: not just for large language models, but text-to-speech, speech-to-text, image generation, video understanding, coding generation and embedding models.
- Production-ready: models are automatically exposed through a single OpenAI-like API and a ChatGPT-like UI playground, with off-the-shelf monitoring and evaluation framework.
- Compatible with most popular model engines
- Easy to expand to custom workloads
Powered by Kalavai
- CoGen AI: A community hosted alternative to OpenAI API for unlimited inference.
- Create your own Free Cursor/Windsurf Clone
Latest updates
- 11 June 2025: Native support for Mac and Raspberry pi devices (ARM).
- 20 February 2025: New shiny GUI interface to control LLM pools and deploy models
- 31 January 2025:
kalavai-clientis now a PyPI package, easier to install than ever!
More news
- 27 January 2025: Support for accessing pools from remote computers
- 9 January 2025: Added support for Aphrodite Engine models
- 8 January 2025: Release of a free, public, shared pool for community LLM deployment
- 24 December 2024: Release of public BOINC pool to donate computing to scientific projects
- 23 December 2024: Release of public petals swarm
- 24 November 2024: Common pools with private user spaces
- 30 October 2024: Release of our public pool platform
Support for AI engines
We currently support out of the box the following AI engines:
- vLLM: most popular GPU-based model inference.
- llama.cpp: CPU-based GGUF model inference.
- SGLang: Super fast GPU-based model inference.
- n8n: no-code workload automation framework.
- Flowise: no-code agentic AI workload framework.
- Speaches: audio (speech-to-text and text-to-speech) model inference.
- Langfuse: open source evaluation and monitoring GenAI framework.
- OpenWebUI: ChatGPT-like UI playground to interface with any models.
Coming soon:
Not what you were looking for? Tell us what engines you'd like to see.
Kalavai is at an early stage of its development. We encourage people to use it and give us feedback! Although we are trying to minimise breaking changes, these may occur until we have a stable version (v1.0).
Want to know more?
- Get a free Kalavai account and access unlimited AI.
- Full documentation for the project.
- Join our Substack for updates and be part of our community
- Join our discord community
Getting started
The kalavai-client is the main tool to interact with the Kalavai platform, to create and manage both local and public pools and also to interact with them (e.g. deploy models). Let's go over its installation.
Requirements
For seed nodes:
- A 64 bits x86 based Linux machine (laptop, desktop or VM)
- Docker engine installed with privilege access.
For workers sharing resources with the pool:
- A laptop, desktop or Virtual Machine (MacOS, Linux or Windows; ARM or x86)
- If self-hosting, workers should be on the same network as the seed node. Looking for over-the-internet connectivity? Check out our managed seeds
- Docker engine installed (for linux, Windows and MacOS) with privilege access.
Install the client
The client is a python package and can be installed with one command:
pip install kalavai-client
Create a a local, private AI pool
You can create and manage your pools with the new kalavai GUI, which can be started with:
kalavai gui start
This will expose the GUI and the backend services in localhost. By default, the GUI is accessible via http://localhost:49153. In the UI users can create and join AI pools, monitor devices, deploy LLMs and more.
Check out our getting started guide for next steps on how to add more workers to your pool, or use our managed seeds service for over-the-internet AI pools.
Enough already, let's run stuff!
For an end to end tour on building your own OpenAI-like service, check our self-hosted guide.
Check our examples to put your new AI pool to good use!
Compatibility matrix
If your system is not currently supported, open an issue and request it. We are expanding this list constantly.
Hardware and OS compatibility
OS compatibility
Currently seed nodes are supported exclusively on linux machines (x86_64 platform). However Kalavai supports mix-pools, i.e. having Windows and MacOS computers as workers.
Since worker nodes run inside docker, any machine that can run docker should be compatible with Kalavai. Here are instructions for linux, Windows and MacOS.
The kalavai client, which controls and access pools, can be installed on any machine that has python 3.10+.
Hardware compatibility:
amd64orx86_64CPU architecture for seed and worker nodes.arm64CPU architecture for worker nodes.- NVIDIA GPU
- Mac M series, AMD and Intel GPUs are currently not supported (interested in helping us test it?)
Roadmap
- Kalavai client on Linux
- [TEMPLATE] Distributed LLM deployment
- Kalavai client on Windows (worker only)
- Kalavai client on Windows WSL2 (seed and worker)
- Self-hosted LLM pools
- Collaborative LLM deployment
- Ray cluster support
- Kalavai client on Mac (worker only)
- Kalavai pools UI
- Support for AMD GPUs
- Support for Mac M GPUs
- Docker install path
Anything missing here? Give us a shout in the discussion board
Contribute
- PR welcome!
- Join the community and share ideas!
- Report bugs, issues and new features.
- Help improve our compatibility matrix by testing on different operative systems.
- Follow our Substack channel for news, guides and more.
- Community integrations are template jobs built by Kalavai and the community that makes deploying distributed workflows easy for users. Anyone can extend them and contribute to the repo.
Star History
Build from source
Expand
Python version >= 3.10.
sudo add-apt-repository ppa:deadsnakes/ppa
sudo apt update
sudo apt install python3.10 python3.10-dev python3-virtualenv python3-venv
virtualenv -p python3.10 env
source env/bin/activate
sudo apt install python3.10-venv python3.10-dev -y
pip install -U setuptools
pip install -e .[dev]
Build python wheels:
bash publish.sh build
Unit tests
To run the unit tests, use:
python -m unittest
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file kalavai_client-0.7.9.tar.gz.
File metadata
- Download URL: kalavai_client-0.7.9.tar.gz
- Upload date:
- Size: 49.4 kB
- Tags: Source
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
51b608626a4dd90ca30f56a146df6c95e580b7225614c2766bbe531e826c50d2
|
|
| MD5 |
c3278d9c4339f234c012c9288c8cd1bf
|
|
| BLAKE2b-256 |
74bc702e8294f689912c6fd36bed8d0b1b3abb567ab23694a7efa4a4d96bf512
|
Provenance
The following attestation bundles were made for kalavai_client-0.7.9.tar.gz:
Publisher:
release.yml on kalavai-net/kalavai-client
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
kalavai_client-0.7.9.tar.gz -
Subject digest:
51b608626a4dd90ca30f56a146df6c95e580b7225614c2766bbe531e826c50d2 - Sigstore transparency entry: 610932868
- Sigstore integration time:
-
Permalink:
kalavai-net/kalavai-client@697a029f50f6df29fe9580c027544a46b222cba9 -
Branch / Tag:
refs/tags/v0.7.9 - Owner: https://github.com/kalavai-net
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
release.yml@697a029f50f6df29fe9580c027544a46b222cba9 -
Trigger Event:
release
-
Statement type:
File details
Details for the file kalavai_client-0.7.9-py3-none-any.whl.
File metadata
- Download URL: kalavai_client-0.7.9-py3-none-any.whl
- Upload date:
- Size: 53.3 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.7
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
1d758d290d57d404849f32d9e7264b30b380b9256eba9074e282b61bd26f9290
|
|
| MD5 |
4c956369774554a676b9ef59a09b1434
|
|
| BLAKE2b-256 |
ed962ffe42ca8982a732880dca097696dfe08ed783aab70b4875f78bb27c4a54
|
Provenance
The following attestation bundles were made for kalavai_client-0.7.9-py3-none-any.whl:
Publisher:
release.yml on kalavai-net/kalavai-client
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
kalavai_client-0.7.9-py3-none-any.whl -
Subject digest:
1d758d290d57d404849f32d9e7264b30b380b9256eba9074e282b61bd26f9290 - Sigstore transparency entry: 610932897
- Sigstore integration time:
-
Permalink:
kalavai-net/kalavai-client@697a029f50f6df29fe9580c027544a46b222cba9 -
Branch / Tag:
refs/tags/v0.7.9 - Owner: https://github.com/kalavai-net
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
release.yml@697a029f50f6df29fe9580c027544a46b222cba9 -
Trigger Event:
release
-
Statement type: