Skip to main content

Finetrainers is a work-in-progress library to support (accessible) training of diffusion models

Project description

finetrainers 🧪

Finetrainers is a work-in-progress library to support (accessible) training of diffusion models and various commonly used training algorithms.

Your browser does not support the video tag. Your browser does not support the video tag.
CogVideoX LoRA training as the first iteration of this project Replication of PikaEffects

Table of Contents

Quickstart

Clone the repository and make sure the requirements are installed: pip install -r requirements.txt and install diffusers from source by pip install git+https://github.com/huggingface/diffusers. The requirements specify diffusers>=0.32.1, but it is always recommended to use the main branch of Diffusers for the latest features and bugfixes. Note that the main branch for finetrainers is also the development branch, and stable support should be expected from the release tags.

Checkout to the latest stable release tag:

git fetch --all --tags
git checkout tags/v0.1.0

Follow the instructions mentioned in the README for the latest stable release.

Using the main branch

To get started quickly with example training scripts on the main development branch, refer to the following:

The following are some simple datasets/HF orgs with good datasets to test training with quickly:

Please checkout docs/models and examples/training to learn more about supported models for training & example reproducible training launch scripts. For a full list of arguments that can be set for training, refer to docs/args.

[!IMPORTANT] It is recommended to use Pytorch 2.5.1 or above for training. Previous versions can lead to completely black videos, OOM errors, or other issues and are not tested. For fully reproducible training, please use the same environment as mentioned in environment.md.

Features

  • DDP, FSDP-2 & HSDP support for all models
  • LoRA and full-rank finetuning; Conditional Control training
  • Memory-efficient single-GPU training
  • Multiple attention backends supported - flash, flex, sage, xformers (see attention docs)
  • Auto-detection of commonly used dataset formats
  • Combined image/video datasets, multiple chainable local/remote datasets, multi-resolution bucketing & more
  • Memory-efficient precomputation support with/without on-the-fly precomputation for large scale datasets
  • Standardized model specification format for training arbitrary models
  • Fake FP8 training (QAT upcoming!)

News

  • 🔥 2025-04-25: Support for different attention providers added!
  • 🔥 2025-04-21: Wan I2V supported added!
  • 🔥 2025-04-12: Channel-concatenated control conditioning support added for CogView4 and Wan!
  • 🔥 2025-04-08: torch.compile support added!
  • 🔥 2025-04-06: Flux support added!
  • 🔥 2025-03-07: CogView4 support added!
  • 🔥 2025-03-03: Wan T2V support added!
  • 🔥 2025-03-03: We have shipped a complete refactor to support multi-backend distributed training, better precomputation handling for big datasets, model specification format (externally usable for training custom models), FSDP & more.
  • 🔥 2025-02-12: We have shipped a set of tooling to curate small and high-quality video datasets for fine-tuning. See video-dataset-scripts documentation page for details!
  • 🔥 2025-02-12: Check out eisneim/ltx_lora_training_i2v_t2v! It builds off of finetrainers to support image to video training for LTX-Video and STG guidance for inference.
  • 🔥 2025-01-15: Support for naive FP8 weight-casting training added! This allows training HunyuanVideo in under 24 GB upto specific resolutions.
  • 🔥 2025-01-13: Support for T2V full-finetuning added! Thanks to @ArEnSc for taking up the initiative!
  • 🔥 2025-01-03: Support for T2V LoRA finetuning of CogVideoX added!
  • 🔥 2024-12-20: Support for T2V LoRA finetuning of Hunyuan Video added! We would like to thank @SHYuanBest for his work on a training script here.
  • 🔥 2024-12-18: Support for T2V LoRA finetuning of LTX Video added!

Support Matrix

The following trainers are currently supported:

[!NOTE] The following numbers were obtained from the release branch. The main branch is unstable at the moment and may use higher memory.

Model Name Tasks Min. LoRA VRAM* Min. Full Finetuning VRAM^
LTX-Video Text-to-Video 5 GB 21 GB
HunyuanVideo Text-to-Video 32 GB OOM
CogVideoX-5b Text-to-Video 18 GB 53 GB
Wan Text-to-Video TODO TODO
CogView4 Text-to-Image TODO TODO
Flux Text-to-Image TODO TODO

*Noted for training-only, no validation, at resolution 49x512x768, rank 128, with pre-computation, using FP8 weights & gradient checkpointing. Pre-computation of conditions and latents may require higher limits (but typically under 16 GB).
^Noted for training-only, no validation, at resolution 49x512x768, with pre-computation, using BF16 weights & gradient checkpointing.

If you would like to use a custom dataset, refer to the dataset preparation guide here.

Featured Projects 🔥

Checkout some amazing projects citing finetrainers:

Checkout the following UIs built for finetrainers:

Acknowledgements

  • finetrainers builds on top of & takes inspiration from great open-source libraries - transformers, accelerate, torchtune, torchtitan, peft, diffusers, bitsandbytes, torchao and deepspeed - to name a few.
  • Some of the design choices of finetrainers were inspired by SimpleTuner. `

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

finetrainers-0.2.0.tar.gz (131.8 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

finetrainers-0.2.0-py3-none-any.whl (176.2 kB view details)

Uploaded Python 3

File details

Details for the file finetrainers-0.2.0.tar.gz.

File metadata

  • Download URL: finetrainers-0.2.0.tar.gz
  • Upload date:
  • Size: 131.8 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.10.14

File hashes

Hashes for finetrainers-0.2.0.tar.gz
Algorithm Hash digest
SHA256 22751c875ad66f7416a5626d79be999c54221f8d1a7822efcdec214124ca8d5f
MD5 66019433f47855dd26d357bb9e4d9753
BLAKE2b-256 1e56939748f7f93e2bf5775afd74e289bcf49803ead7344ef1f3027c46e160c6

See more details on using hashes here.

File details

Details for the file finetrainers-0.2.0-py3-none-any.whl.

File metadata

  • Download URL: finetrainers-0.2.0-py3-none-any.whl
  • Upload date:
  • Size: 176.2 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.10.14

File hashes

Hashes for finetrainers-0.2.0-py3-none-any.whl
Algorithm Hash digest
SHA256 ca33b5fb1dcd559e26a9a43d11923fe3ebe9221ebaf553e9e6b24862eaca8765
MD5 1ecbd8950d69820c88f4dd09530f10ca
BLAKE2b-256 39200c5d6e749511b29a57a852e79f83ad4efc9637310aee78f26c68bd67c76e

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page