Skip to main content

Lookahead Keys Attention

Project description

Lookahead Keys Attention (wip)

Causal Attention with Lookahead Keys

Installation

pip install lookahead-keys-attention

Usage

import torch
from lookahead_keys_attention import Castle

# Initialize the Castle attention module
model = Castle(
    dim=512,           # input dimension
    heads=8,           # number of attention heads
    dim_head=64,       # dimension per head
    use_triton=None    # auto-detect CUDA for Triton optimization
)

# Example with CUDA sequence
batch_size = 2
seq_len = 128
dim = 512

# Move to CUDA if available
device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')
model = model.to(device)

# Input sequence
x = torch.randn(batch_size, seq_len, dim).to(device)

# Forward pass
output = model(x)  # Shape: [batch_size, seq_len, dim]

# For inference with caching (single token generation)
cache = None
for i in range(seq_len):
    token = x[:, i:i+1, :]  # Single token
    output, cache = model(token, cache=cache, return_next_cache=True)

Citations

@inproceedings{Song2025CausalAW,
    title   = {Causal Attention with Lookahead Keys},
    author  = {Zhuoqing Song and Peng Sun and Huizhuo Yuan and Quanquan Gu},
    year    = {2025},
    url     = {https://api.semanticscholar.org/CorpusID:281218151}
}

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

lookahead_keys_attention-0.0.5.tar.gz (154.0 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

lookahead_keys_attention-0.0.5-py3-none-any.whl (10.6 kB view details)

Uploaded Python 3

File details

Details for the file lookahead_keys_attention-0.0.5.tar.gz.

File metadata

  • Download URL: lookahead_keys_attention-0.0.5.tar.gz
  • Upload date:
  • Size: 154.0 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.9.23

File hashes

Hashes for lookahead_keys_attention-0.0.5.tar.gz
Algorithm Hash digest
SHA256 ff65f52041204080466704e01782547140c3728c4405b34b4a7f615a45f89ad1
MD5 ba324c61c2f5e0baf6b5fc5d201234b2
BLAKE2b-256 7b9c66cf1c34c796664456799f7596cc5612bc47b59fdf1ec2f0fe3988912004

See more details on using hashes here.

File details

Details for the file lookahead_keys_attention-0.0.5-py3-none-any.whl.

File metadata

File hashes

Hashes for lookahead_keys_attention-0.0.5-py3-none-any.whl
Algorithm Hash digest
SHA256 aae1df41ebecaadf6d078b21e95ca84e6d805e8a128a8d71e69223aec3b4191d
MD5 92e0a2af8f0871aef7a87942ca3c0560
BLAKE2b-256 a309c0051b45b445f7630f45dbe357753e02b4eb6b5414eecb3c2165b80e01e7

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page