CUDA and Triton implementations of Flash Attention with SoftmaxN.
Project description
The author of this package has not provided a project description
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Close
Hashes for flash-attention-softmax-n-0.1.0.tar.gz
Algorithm | Hash digest | |
---|---|---|
SHA256 | ea8aa79fb5b3e8d27d26ee1109d4e2e7691af4b95fe1e4fa50bf3ac56492b465 |
|
MD5 | c221862e6915d91991221b983e36d243 |
|
BLAKE2b-256 | 0e4fcc434c22d8d6306db3c5cb3d57bb81936e7e1af6a30c33d16935493ded5b |
Close
Hashes for flash_attention_softmax_n-0.1.0-py3-none-any.whl
Algorithm | Hash digest | |
---|---|---|
SHA256 | c8ad2259ff8ad18e341c089a6acbedee18a107245e21ad66275df0db967d75b2 |
|
MD5 | 9ad97518b78ec387c74e8c56e2e3fed8 |
|
BLAKE2b-256 | 03e452a47eaff9cce0e4c9d5dcf801a44fd9f2abee05dfba50f8a446eda6265b |