Benchmark-backed Metal Flash Attention backends for MLX on Apple Silicon
Project description
mlx-mfa
mlx-mfa is a Metal Flash Attention + serving-oriented runtime layer for MLX on
Apple Silicon. It provides high-performance attention kernels, runtime helpers,
and cache abstractions for dense training/inference plus modern serving flows.
Current version: 2.27.0 — native attn_bias Metal kernel + DiT/UNet dispatch audit.
Foreword
MLX Metal Flash Attention - Why?
I've been working on personal ports of Video Super Resolution and Video Reconstruction models for months, but always ended up frustrated by the slow inference in my M1 Max MacBook Pro. And to try to mitigate this without having to buy a brand-new, very expensive new M4, then M5 Max, I decided to at least try to port Flash Attention to Mac, hoping for better results. And having better results porting VSR/VR models to MLX than MPS, that's why I ended up doing it.
At this point, despite the lower than hoped for results, I'm still pretty satisfied with the results in my M1 Max MBP.
I'll be doing only reduced work on this project until June 2026, when I'll upgrade from my M1 Max to a M5 Max MBP, with which I expect to be able to obtain much better results, thanks to the improvements Apple has been adding to its silicon.
v2.27.0 adds native Metal attn_bias kernel support (additive bias on
attention logits without SDPA fallback), a dispatch audit for 11 DiT/UNet
architectures, and varlen validation for token merging workflows.
See CHANGELOG.md for full details per version.
Thank you for your interest, and let me know if you've been able to improve on my work!
Current Repository Status
- V2 dense is the main production path.
- Strongest dense wins on M1 Max remain causal D=64/128 and tile-skip regimes (window/sparse).
- D=256 is narrow benchmark-backed only (not broad promotion).
- D=512 remains SDPA-default.
- Native dense backward was benchmarked and not promoted.
- Sage is a specialized decode backend (narrow, benchmark-gated use).
- V3/V4/V5 remain experimental/hardware-dependent.
- TurboQuant KV cache compression (Phase 1–4) production-ready.
- SVDQuantLinear W4A16 + optional SVD low-rank correction for DiT quantization.
- GNA native kernel inline 3D window attention (D=128, f16/bf16, forward-only).
- Native
attn_biasadditive bias on logits via Metal kernel (modes 1/2: per-KV and per-head per-KV broadcast). - Serving/runtime capability surface is now substantially expanded:
- paged KV + packed varlen query support
- paged continuous batching/remap
- explicit chunked prefill
- runtime-managed prefix reuse
- runtime speculative draft/verify flow
- deeper splitfuse runtime integration
- KV cache abstraction layer
- minimal real hybrid/offload-capable cache behavior (local offload tier)
- TurboQuant compressed KV serving (
create_decode_runtime(turboquant=True))
Limitations
- Main validation hardware is Apple M1 Max.
- Broad parity claims against CUDA FlashAttention ecosystems are not made.
- Some advanced paths are intentionally narrow, bridge-based, or explicit-only.
- Hybrid offload is currently a local offload milestone, not remote/ distributed cache infrastructure.
- Future major hardware-specific optimization work is deferred pending newer Apple hardware (M5+).
Best M1 Max Benchmark Highlights
Representative benchmark-backed outcomes (see RESULTS.md and
docs/benchmarks/RESULTS.md for details):
| Area | Representative result (M1 Max) | Interpretation |
|---|---|---|
| Dense causal V2 | up to ~1.82x vs SDPA (D=64, N=8192) | Primary production win regime |
| Dense causal V2 | up to ~1.75x vs SDPA (D=128, N=16384) | Strong long-sequence causal performance |
| Sliding window | up to ~21x vs full SDPA | Tile-skip regime remains strongest |
| D=256 | narrow causal long-N wins (for example ~1.16x at N=16384 f16) | Keep narrow policy only |
| D=512 | decision pass found no broad wins | SDPA-default remains correct |
Serving/Runtime Capability Summary
| Capability | Maturity | Current status |
|---|---|---|
| Paged KV decode runtime | Fully usable | Explicit runtime/API usage; no broad auto-promotion |
| Paged + packed varlen queries | Production (fused kernel) | Single-dispatch fused kernel for all query/KV length combinations |
| Paged continuous batching remap | Fully usable | Explicit cache_batch_idx semantics + runtime helpers |
| Chunked prefill | Fully usable (scheduler-oriented) | Operational capability; not a throughput win on current matrix |
| Runtime prefix caching | Fully usable | Register/seed/reuse path integrated with runtime metadata |
| Runtime speculative decode | Fully usable (narrow) | speculative_step + verify integration; scheduler engine still future work |
| Splitfuse runtime integration | Narrow/conditional | Runtime path exists; performance remains shape-sensitive |
| Hybrid KV cache + local offload tier | Narrow/conditional milestone | Real hot/cold/offloaded behavior locally; remote offload future work |
| TurboQuant KV compression (Phase 4) | Production | 5.33× K compression, WHT fused in kernel (1.1–1.4× faster) |
| SVDQuantLinear | Production | W4A16 + rank-r FP16 correction; quantize_model() tree walker |
| GNA native kernel | Production | Inline 3D window attention (D=128); exact per-element masking |
Native attn_bias |
Production | Modes 1/2 via V2 STEEL; modes 0/3 SDPA fallback |
| External cache adapter layer | Experimental groundwork | Concrete local backend provided; external backend integrations pending |
Repository Guide
- Feature coverage:
docs/FEATURE_COVERAGE.md - API manual:
docs/API_MANUAL.md - Architecture:
docs/ARCHITECTURE.md - Inventory map:
docs/INVENTORY.md - Benchmark interpretation:
docs/benchmarks/RESULTS.md - Root benchmark summary:
RESULTS.md - Changelog:
CHANGELOG.md - Historical development archive:
devnotes/ - Examples:
examples/
Production vs Narrow vs Experimental
| Status | Components |
|---|---|
| Production | V2 dense causal small-D path; window/sparse tile-skip; SDPA fallback policy; TurboQuant KV compression; SVDQuantLinear; GNA native kernel; native attn_bias |
| Narrow / conditional | D=256 causal long-N policy; Sage decode regimes; splitfuse/page-native runtime paths; hybrid local offload behavior |
| Experimental | V3/V4/V5 families; external/LMCache-like backend extensions beyond local adapter |
Recommended Usage
- Use
backend="auto"for dense attention and let policy route between V2 and SDPA. - Use
create_decode_runtime(...)for serving flows instead of stitching helper calls manually. - Treat paged/packed/chunked/prefix/speculative features as explicit runtime capabilities.
- Use Sage as a specialized decode backend only when your workload matches the benchmark-backed regime.
Installation
pip install -e .
Minimal Usage
import mlx.core as mx
from mlx_mfa import flash_attention, flash_attention_gna, create_decode_runtime
from mlx_mfa import SVDQuantLinear, quantize_model
# Dense attention
q = mx.random.normal((1, 8, 1024, 128)).astype(mx.float16)
k = mx.random.normal((1, 8, 1024, 128)).astype(mx.float16)
v = mx.random.normal((1, 8, 1024, 128)).astype(mx.float16)
out = flash_attention(q, k, v, causal=True)
# Token merging proportional attention (native Metal, no SDPA fallback)
merge_counts = mx.ones((1, 1, 1, 1024), dtype=mx.float16)
merge_counts[..., :256] = 2.0 # first 256 tokens are merged pairs
bias = mx.log(merge_counts) # [1, 1, 1, N_kv] — mode 1 broadcast
out_biased = flash_attention(q, k, v, attn_bias=bias)
# GNA (Generalized Neighborhood Attention) — 3D window
# Video: 8 frames of 32x32, local 3D window, sliding
q_vid = mx.random.normal((1, 8, 8192, 128)).astype(mx.float16)
k_vid = mx.random.normal((1, 8, 8192, 128)).astype(mx.float16)
v_vid = mx.random.normal((1, 8, 8192, 128)).astype(mx.float16)
out_gna = flash_attention_gna(q_vid, k_vid, v_vid,
seq_shape=(8, 32, 32),
window_size=(2, 8, 8),
stride=(1, 1, 1))
# SVDQuantLinear — W4A16 + SVD low-rank correction
# (quantize_model replaces nn.Linear layers in-place)
# model = quantize_model(model, group_size=64, bits=4, rank=32)
# Serving-oriented runtime
rt = create_decode_runtime(
backend="auto",
paged=False,
quantized_kv=False,
B=1,
H_q=8,
H_kv=8,
D=128,
max_seq_len=4096,
)
out_prefill = rt.prefill(q, k, v)
out_step = rt.step(
mx.random.normal((1, 8, 1, 128)).astype(mx.float16),
mx.random.normal((1, 8, 1, 128)).astype(mx.float16),
mx.random.normal((1, 8, 1, 128)).astype(mx.float16),
)
License
MIT. See LICENSE.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file mlx_mfa-2.27.0.tar.gz.
File metadata
- Download URL: mlx_mfa-2.27.0.tar.gz
- Upload date:
- Size: 740.5 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.11.14
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
acf05427907236726c2470d362f992af79aa6f6634dc8b0c17f1f501fd7f25be
|
|
| MD5 |
0faf07cf4bea5fe60d13bd8f8d01dbdb
|
|
| BLAKE2b-256 |
0d0cdb565376b7434cbdb51426a56e19e5886a3009219aed519acb491527e71e
|
File details
Details for the file mlx_mfa-2.27.0-cp311-cp311-macosx_26_0_arm64.whl.
File metadata
- Download URL: mlx_mfa-2.27.0-cp311-cp311-macosx_26_0_arm64.whl
- Upload date:
- Size: 408.0 kB
- Tags: CPython 3.11, macOS 26.0+ ARM64
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.11.14
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
1df988103c8da71046ef6d3dc2ac4a5f41f98a85fbde2f70e2f37c2cd3e9b1d5
|
|
| MD5 |
fdc0d2317d112a11bee08f1734aee3bb
|
|
| BLAKE2b-256 |
0dd4e8477799d1bf44fdec1abd478c60dad455d5ea9a63f8958dd9689bbf537d
|