FMS Acceleration Plugin for Functionalities Used in Instruct Lab Training
Project description
FMS Acceleration for Instruct Lab
This library contains plugins to accelerate finetuning with the following optimizations:
- Padding-Free Flash Attention Computation
Plugins
Plugin | Description | Depends | Loading | Augmentation | Callbacks |
---|---|---|---|---|---|
padding_free | Padding-Free Flash Attention Computation | flash_attn | ✅ | ✅ |
Native Transformers Support from v4.44.0
Transformers natively supports padding-free from v4.44.0 see here. The padding-free plugin will use the transformers library if compatible,
otherwise if transformers < v4.44.0
the plugin will use an internal implementation instead.
Known Issues
Currently Only Supports Pre-Tokenized Dataset
The padding-free plugin currently only works with pre-tokenized datasets, this is because it is currently designed to replace
the data collator from SFTTrainer
with a custom data collator to manipulate the input to the modified flash attention forward.
There are some cases, the data collator for SFTTrainer will handle the formatting and tokenization from raw text datasets. The plugin is currently unable to both handle the original data collation and apply its custom data collator over it as the same time. This issue will be addressed in a future commit to support this case.
In the meantime, the plugin expects the user to provide a pretokenized dataset that
- is formatted with a template for instruct-tuning cases
- is tokenized
- has template labels that are masked to exclude from loss computation
- has eos token appended
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distributions
Built Distribution
File details
Details for the file fms_acceleration_ilab-0.1.0-py3-none-any.whl
.
File metadata
- Download URL: fms_acceleration_ilab-0.1.0-py3-none-any.whl
- Upload date:
- Size: 8.7 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/5.1.0 CPython/3.12.4
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 376f84b064990cf647863c29cb3037f0a47cf4a2a9e5cecb954110a33ccae25c |
|
MD5 | 5a4f93080a595bb1a9c7b3ff1388eb37 |
|
BLAKE2b-256 | a02eff21241f823dff4451e55f595a19f4793f8cf262b60c9bc02ce4863d518b |