Skip to main content

Optimized version of llama3 / gemma.

Project description

llaminate

Optimized version of llama3, using tokun.

Neural tokenization

This project is a showcase for a neural tokenization technique. Since the inputs are compressed and have a smaller shape, the LLM is downsized accordingly.

For example, llama3-8b is brought down to 34 million parameters instead of 8 billion.

Installation

Usage

Resources

Models

Notebooks

Final model:

TODO

See TODO.

Credits

This project winks at llama3 from Meta, but doesn't actually its weights nor code.

License

Licensed under the aGPLv3.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

llaminate-0.4.1.tar.gz (4.4 kB view hashes)

Uploaded Source

Built Distribution

llaminate-0.4.1-py3-none-any.whl (6.1 kB view hashes)

Uploaded Python 3

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page