prepare your dataset for finetuning LLMs
Project description
Dataset Preparation for Transformers Fine-tuning
The Dataset Prep Transformers package simplifies the process of preparing datasets for fine-tuning or training various large language models available in the Hugging Face Transformers library. Whether you're using a model from the Hugging Face repository or have your own dataset, this package streamlines the data integration for a seamless training experience.
Features
- Easily integrate your dataset with Hugging Face Transformers models for training or fine-tuning.
- Specify the model repository ID and dataset from the Hugging Face library to automatically fetch and configure the data.
- Seamlessly incorporate your custom dataset by providing it as input to the package.
- new: custom map function => map_function = your_function (having the facility of tokenization)
Installation
You can install the package using pip:
pip install yDataPrep
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
yDataPrep-0.0.1.2.1.tar.gz
(3.1 kB
view details)
Built Distribution
File details
Details for the file yDataPrep-0.0.1.2.1.tar.gz
.
File metadata
- Download URL: yDataPrep-0.0.1.2.1.tar.gz
- Upload date:
- Size: 3.1 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.2 CPython/3.11.5
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 335ce0184ad133ad4720bd28001715543c3293e1af102ed0498909f69bba71b5 |
|
MD5 | 921617b92b96ae20e9f26a45393cb708 |
|
BLAKE2b-256 | 80d5905ad62e31394f02715fa3ccc23b395c869692e5f906382adf3ba84e8f47 |
File details
Details for the file yDataPrep-0.0.1.2.1-py3-none-any.whl
.
File metadata
- Download URL: yDataPrep-0.0.1.2.1-py3-none-any.whl
- Upload date:
- Size: 3.3 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.2 CPython/3.11.5
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 524c3c7bd430aab544da8608bac69ba75b0733f7f15fb3a1fa17e917aad53f14 |
|
MD5 | 7fc986502237bf811b358395b9af06ac |
|
BLAKE2b-256 | 7b668b1105e53188384e3f967c0d62580bbce53e6f0da27f9426187f99227d69 |