Skip to main content

Fast Weight Attention

Project description

Fast Weight Attention (wip)

An attention based fast weight episodic memory, in the same vein as the memory MLP from TTT / Titans and fwPKM from Sakana AI

Citations

@article{zhang2026loger,
    title   = {LoGeR: Long-Context Geometric Reconstruction with Hybrid Memory},
    author  = {Zhang, Junyi and Herrmann, Charles and Hur, Junhwa and Sun, Chen and Yang, Ming-Hsuan and Cole, Forrester and Darrell, Trevor and Sun, Deqing},
    journal = {arXiv preprint arXiv:2603.03269},
    year    = {2026}
}
@misc{zhao2026fastweightproductkeymemory,
    title   = {Fast-weight Product Key Memory},
    author  = {Tianyu Zhao and Llion Jones},
    year    = {2026},
    eprint  = {2601.00671},
    archivePrefix = {arXiv},
    primaryClass = {cs.CL},
    url     = {https://arxiv.org/abs/2601.00671},
}
@misc{jordan2024muon,
    author       = {Keller Jordan and Yuchen Jin and Vlado Boza and Jiacheng You and Franz Cesista and Laker Newhouse and Jeremy Bernstein},
    title        = {Muon: An optimizer for hidden layers in neural networks},
    year         = {2024},
    url          = {https://kellerjordan.github.io/posts/muon/}
}

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

fast_weight_attention-0.0.1.tar.gz (6.4 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

fast_weight_attention-0.0.1-py3-none-any.whl (5.8 kB view details)

Uploaded Python 3

File details

Details for the file fast_weight_attention-0.0.1.tar.gz.

File metadata

File hashes

Hashes for fast_weight_attention-0.0.1.tar.gz
Algorithm Hash digest
SHA256 a32d7a14aefc524e51297c896d87b950add5eb3409a7bb6f63aaef4564fd3ba2
MD5 1727357b124910719476c08700680662
BLAKE2b-256 9414e1253ef1a68a370c9e7d9c906a6e88dd8057a64842cbf71339300d12eeff

See more details on using hashes here.

File details

Details for the file fast_weight_attention-0.0.1-py3-none-any.whl.

File metadata

File hashes

Hashes for fast_weight_attention-0.0.1-py3-none-any.whl
Algorithm Hash digest
SHA256 15dd2dcada37c26fa859c30a3f5111e2eb3ca4166b9a6fcf1b649eb411162a45
MD5 f7ec7ba0f959250f2bb324b3ae4bda2c
BLAKE2b-256 d81370f131102faf19508426fbf5dfde70db6f13fd3f1417865c326f3bf3a256

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page