AI Handler: An engine which wraps certain huggingface models
Project description
AI Handler
This is a simple framework for running AI models. It makes use of the huggingface API which gives you a queue, threading, a simple API, and the ability to run Stable Diffusion and LLMs seamlessly from your local hardware.
This is not intended to be used as a standalone application.
It can easily be extended and used to power interfaces or it can be run from the command line.
AI Handler is a work in progress. It powers two projects at the moment, but may not be ready for general use.
Installation
This is a work in progress.
Pre-requisites
System requirements
- Windows 10+
- Python 3.10.8
- pip 23.0.1
- CUDA toolkit 11.7
- CUDNN 8.6.0.163
- Cuda capable GPU
- 16gb+ ram
For Windows, follow windows branch instructions
Install
pip install https://github.com/w4ffl35/diffusers/archive/refs/tags/v0.15.0.ckpt_fix_0.0.1.tar.gz
pip install aihandler
Optional
These are optional instructions for installing TensorRT and Deepspeed for Windows
Install Tensor RT:
- Download TensorRT-8.4.3.1.Windows10.x86_64.cuda-11.6.cudnn8.4
- Git clone TensorRT 8.4.3.1
- Follow their instructions to build TensorRT-8.4.3.1 python wheel
- Install TensorRT
pip install tensorrt-*.whl
Install Deepspeed:
- Git clone Deepspeed 0.8.1
- Follow their instructions to build Deepspeed python wheel
- Install Deepspeed `pip install deepspeed-*.whl
Environment variables
AIRUNNER_ENVIRONMENT
-dev
orprod
. Defaults todev
. This controls the LOG_LEVELLOG_LEVEL
-FATAL
for production,DEBUG
for development. Override this to force a log level
Huggingface variables
Offline mode
These environment variables keep you offline until you need to download a model. This prevents unwanted online access and speeds up usage of huggingface libraries.
DISABLE_TELEMETRY
Keep this set to 1 at all times. Huggingface collects minimal telemetry when downloading a model from their repository but this will keep it disabled. See more info in this github threadHF_HUB_OFFLINE
When loading a diffusers model, huggingface libraries will attempt to download an updated cache before running the model. This prevents that check from happening (long with a boolean passed toload_pretrained
see the runner.py file for examples)TRANSFORMERS_OFFLINE
Similar toHF_HUB_OFFLINE
but for transformers models
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file aihandler-1.19.0.tar.gz
.
File metadata
- Download URL: aihandler-1.19.0.tar.gz
- Upload date:
- Size: 54.5 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.2 CPython/3.9.17
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 09cab787b5b54bd8d2a6f7c475c5693ba3e0f21d94a2261734ef66450ca4f829 |
|
MD5 | 975a991022d7be575f522c913d092683 |
|
BLAKE2b-256 | 17c9b6a9cdfa9389b7a7989cd3bfa284c8ada70e972668f2d7910ef31addee95 |
File details
Details for the file aihandler-1.19.0-py3-none-any.whl
.
File metadata
- Download URL: aihandler-1.19.0-py3-none-any.whl
- Upload date:
- Size: 61.3 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.2 CPython/3.9.17
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 5f2c434a9f13e51fedd4639df4bb5717180eb15a8dfb2bf2d3f59b728ff5af19 |
|
MD5 | ac0e49c824d53709d1a0f5b907d24260 |
|
BLAKE2b-256 | 7e2c3f79edb53ddd7906ef53a9ab26267fb3aa5bbb8993c162ad4536d6ecaf06 |