A hybrid chatbot - LLM side.
Project description
WAFL-llm
WAFL is built to run as a two-part system. Both can be installed on the same machine. This is the LLM side of the WAFL project.
LLM side (needs a GPU)
This is a model server for the speech-to-text model, the LLM, the embedding system, and the text-to-speech model.
Installation
In order to quickly run the LLM side, you can use the following installation commands:
pip install wafl-llm
wafl-llm start
which will use the default models and start the server on port 8080.
Docker
A docker image can be used to run it as in the following:
$ docker run -p8080:8080 --env NVIDIA_DISABLE_REQUIRE=1 --gpus all fractalego/wafl-llm:latest
or one can clone this repository and run the following
docker/build.sh
docker run -p8080:8080 --env NVIDIA_DISABLE_REQUIRE=1 --gpus all wafl-llm
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
wafl_llm-0.0.80.tar.gz
(6.9 kB
view hashes)
Built Distribution
Close
Hashes for wafl_llm-0.0.80-py3-none-any.whl
Algorithm | Hash digest | |
---|---|---|
SHA256 | 840cb9943ffcacc08c740765c417ca45acc1b5f210bf5fdddee56d353cc474b8 |
|
MD5 | c4aeeada00bcd85b573ff7483e7f56ef |
|
BLAKE2b-256 | 5fc0567c44c721cb3c941f83d252aa767c6221fb9b87cde7612f89d96a715c57 |