Transformers at zeta scales
Project description
Zeta - A Library for Zetascale Transformers
Docs for Zeta.
Zeta is a PyTorch-powered library, forged in the heart of the Halo array, that empowers researchers and developers to scale up Transformers efficiently and effectively. It leverages seminal research advancements to enhance the generality, capability, and stability of scaling Transformers while optimizing training efficiency.
Installation
To install:
pip install zetascale
To get hands-on and develop it locally:
git clone https://github.com/kyegomez/zeta.git
cd zeta
pip install -e .
Initiating Your Journey
Creating a model empowered with the aforementioned breakthrough research features is a breeze. Here's how to quickly materialize a BERT-like encoder:
>>> from zeta import EncoderConfig
>>> from zeta import Encoder
>>> config = EncoderConfig(vocab_size=64000)
>>> model = Encoder(config)
>>> print(model)
Acknowledgments
Zeta is a masterpiece inspired by elements of FairSeq and UniLM.
Citations
If our work here in Zeta has aided you in your journey, please consider acknowledging our efforts in your work. You can find relevant citation details in our Citations Document.
Contributing
We're always thrilled to welcome new ideas and improvements from the community. Please check our Contributor's Guide for more details about contributing.
- Create an modular omni-universal Attention class with flash multihead attention or regular mh or dilated attention -> then integrate into Decoder/ DecoderConfig
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Hashes for zetascale-0.0.7-py3-none-any.whl
Algorithm | Hash digest | |
---|---|---|
SHA256 | 7bd8d76dd4aad2476ba970f14c513a189b711a2346ddc03af2829d39011521f0 |
|
MD5 | 7a105524aa03e7677d3b1e336fb87f72 |
|
BLAKE2b-256 | 21979210e5a44f53c9b486d7835fa8fcd525a47118053d9c503ba64dc9041559 |