CUDA and Triton implementations of Flash Attention with SoftmaxN.
Project description
The author of this package has not provided a project description
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Close
Hashes for flash-attention-softmax-n-0.1.2.tar.gz
Algorithm | Hash digest | |
---|---|---|
SHA256 | f27feebe1b02d410d4bc5a2f1736168a2bbf77b5359b34ed029b0bcd2d8a61ce |
|
MD5 | ada82ce48584bfe167aca989d441e020 |
|
BLAKE2b-256 | d00989d6f11f17833d20af896173389adb89799f7350f508118a32ea7957959a |
Close
Hashes for flash_attention_softmax_n-0.1.2-py3-none-any.whl
Algorithm | Hash digest | |
---|---|---|
SHA256 | 4df6b062a0c1324d403289e9377518122e6769a370d23f97fc476dbf2db268a2 |
|
MD5 | 9c12ea0f11f2cd3ffc35a3bc3dcc4a55 |
|
BLAKE2b-256 | d7213139d7409eb50735b4c5b3d28526352e430a2a2ebd045ecc45bd33293f3e |