prepare your dataset for finetuning LLMs
Project description
Dataset Preparation for Transformers Fine-tuning
The Dataset Prep Transformers package simplifies the process of preparing datasets for fine-tuning or training various large language models available in the Hugging Face Transformers library. Whether you're using a model from the Hugging Face repository or have your own dataset, this package streamlines the data integration for a seamless training experience.
Features
- Easily integrate your dataset with Hugging Face Transformers models for training or fine-tuning.
- Specify the model repository ID and dataset from the Hugging Face library to automatically fetch and configure the data.
- Seamlessly incorporate your custom dataset by providing it as input to the package.
Installation
You can install the package using pip:
pip install d4train
Project details
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
d4train-0.0.2.2.tar.gz
(3.0 kB
view hashes)
Built Distribution
Close
Hashes for d4train-0.0.2.2-py3-none-any.whl
Algorithm | Hash digest | |
---|---|---|
SHA256 | fc14ba4a55ce6b8455e69896544077e79588c57f053d79ad403a78eb7959b0ca |
|
MD5 | 82e127a4efa63f96f4c3677056efd7ec |
|
BLAKE2b-256 | 180570590d6cfc1a51d6750cddf5ee7730f5ca261f0d5eba75f72214cd1b5aaf |