Skip to main content

Transformers at zeta scales

Project description

Zeta - A Transgalactic Library for Scalable Transformations

MIT License MIT License

Zeta is a PyTorch-powered library, forged in the heart of the Halo array, that empowers researchers and developers to scale up Transformers efficiently and effectively. It leverages seminal research advancements to enhance the generality, capability, and stability of scaling Transformers while optimizing training efficiency.

Installation

To install:

pip install zetascale

To get hands-on and develop it locally:

git clone https://github.com/kyegomez/zeta.git
cd zeta
pip install -e .

Initiating Your Journey

Creating a model empowered with the aforementioned breakthrough research features is a breeze. Here's how to quickly materialize a BERT-like encoder:

>>> from zeta import EncoderConfig
>>> from zeta import Encoder

>>> config = EncoderConfig(vocab_size=64000)
>>> model = Encoder(config)

>>> print(model)

Additionally, we support the Decoder and EncoderDecoder architectures:

# To create a decoder model
>>> from zeta import DecoderConfig
>>> from zeta import Decoder

>>> config = DecoderConfig(vocab_size=64000)
>>> decoder = Decoder(config)
>>> print(decoder)

# To create an encoder-decoder model
>>> from zeta import EncoderDecoderConfig
>>> from zeta import EncoderDecoder

>>> config = EncoderDecoderConfig(vocab_size=64000)
>>> encdec = EncoderDecoder(config)
>>> print(encdec)

Key Features

Most of the transformative features mentioned below can be enabled by simply setting the corresponding parameters in the config:

>>> from zeta import EncoderConfig
>>> from zeta import Encoder

>>> config = EncoderConfig(vocab_size=64000, deepnorm=True, multiway=True)
>>> model = Encoder(config)

>>> print(model)

For a complete overview of our key features, refer to our Feature Guide.

Examples

Discover how to wield Zeta in a multitude of scenarios/tasks, including but not limited to:

We are working tirelessly to expand the collection of examples spanning various tasks (e.g., vision pretraining, speech recognition) and various deep learning frameworks (e.g., DeepSpeed, Megatron-LM). Your comments, suggestions, or contributions are welcome!

Results

Check out our Results Page to witness Zeta's exceptional performance in Stability Evaluations and Scaling-up Experiments.

Acknowledgments

Zeta is a masterpiece inspired by elements of FairSeq and UniLM.

Citations

If our work here in Zeta has aided you in your journey, please consider acknowledging our efforts in your work. You can find relevant citation details in our Citations Document.

Contributing

We're always thrilled to welcome new ideas and improvements from the community. Please check our Contributor's Guide for more details about contributing.

  • Create an modular omni-universal Attention class with flash multihead attention or regular mh or dilated attention -> then integrate into Decoder/ DecoderConfig

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

zetascale-0.0.4.tar.gz (64.0 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

zetascale-0.0.4-py3-none-any.whl (75.4 kB view details)

Uploaded Python 3

File details

Details for the file zetascale-0.0.4.tar.gz.

File metadata

  • Download URL: zetascale-0.0.4.tar.gz
  • Upload date:
  • Size: 64.0 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.3.2 CPython/3.11.0 Darwin/22.4.0

File hashes

Hashes for zetascale-0.0.4.tar.gz
Algorithm Hash digest
SHA256 6fbb0335ed69bd0f6dcb992d9feabc6559b01600d3f0b29b34b34e3a785c70ff
MD5 7be9425fddaf21f4a0c9c716ff853948
BLAKE2b-256 10e3b908bedeed3b586e1dab1269e76aea8d88f0f5b7575025df7874a1e504a5

See more details on using hashes here.

File details

Details for the file zetascale-0.0.4-py3-none-any.whl.

File metadata

  • Download URL: zetascale-0.0.4-py3-none-any.whl
  • Upload date:
  • Size: 75.4 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.3.2 CPython/3.11.0 Darwin/22.4.0

File hashes

Hashes for zetascale-0.0.4-py3-none-any.whl
Algorithm Hash digest
SHA256 dfe5eeaf059cf7c27e447f62896999fd31520bda156ae06cb0f06a578dee41bc
MD5 445bb13a973346a74a63264599e3fa95
BLAKE2b-256 4f2a145f0ccde054b1fd20e650ea9d1f5bcea4c9a27dd7b877025ce305346208

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page