CUDA and Triton implementations of Flash Attention with SoftmaxN.
Project description
The author of this package has not provided a project description
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Close
Hashes for flash-attention-softmax-n-0.1.1.tar.gz
Algorithm | Hash digest | |
---|---|---|
SHA256 | d77e2d832512c569e9341b72fd41bcb904a4a7ae05a2059c904ea994d55511f0 |
|
MD5 | 99120bfd1be848ba44af5a895ebb00d2 |
|
BLAKE2b-256 | eb4063a4d7523c8f29a321b8937149345b9506418776c527ede979c309f45f1f |
Close
Hashes for flash_attention_softmax_n-0.1.1-py3-none-any.whl
Algorithm | Hash digest | |
---|---|---|
SHA256 | ad3c9299f86a2623c9786857fee686a009b817649d7efef3aec0d72e47f1bb63 |
|
MD5 | 872fff2b0a492bd3f7463cb9106387d3 |
|
BLAKE2b-256 | f6b885081626f879c4d5be9fe8dbf3d7a17682779c7e7b6e3bc7160ad6246bff |