Skip to main content

Lookahead Keys Attention

Project description

Lookahead Keys Attention (wip)

Causal Attention with Lookahead Keys

Installation

pip install lookahead-keys-attention

Usage

import torch
from lookahead_keys_attention import Castle

# Initialize the Castle attention module
model = Castle(
    dim=512,           # input dimension
    heads=8,           # number of attention heads
    dim_head=64,       # dimension per head
    use_triton=None    # auto-detect CUDA for Triton optimization
)

# Example with CUDA sequence
batch_size = 2
seq_len = 128
dim = 512

# Move to CUDA if available
device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')
model = model.to(device)

# Input sequence
x = torch.randn(batch_size, seq_len, dim).to(device)

# Forward pass
output = model(x)  # Shape: [batch_size, seq_len, dim]

# For inference with caching (single token generation)
cache = None
for i in range(seq_len):
    token = x[:, i:i+1, :]  # Single token
    output, cache = model(token, cache=cache, return_next_cache=True)

Citations

@inproceedings{Song2025CausalAW,
    title   = {Causal Attention with Lookahead Keys},
    author  = {Zhuoqing Song and Peng Sun and Huizhuo Yuan and Quanquan Gu},
    year    = {2025},
    url     = {https://api.semanticscholar.org/CorpusID:281218151}
}

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

lookahead_keys_attention-0.0.1.tar.gz (153.6 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

lookahead_keys_attention-0.0.1-py3-none-any.whl (10.7 kB view details)

Uploaded Python 3

File details

Details for the file lookahead_keys_attention-0.0.1.tar.gz.

File metadata

  • Download URL: lookahead_keys_attention-0.0.1.tar.gz
  • Upload date:
  • Size: 153.6 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.9.23

File hashes

Hashes for lookahead_keys_attention-0.0.1.tar.gz
Algorithm Hash digest
SHA256 ffd665b5c77769d96bea4c5c5c0048e7b46414e75eff85f82bf092e2a98cff5d
MD5 8b83d59d58f50fbd23bc4cbea93b078e
BLAKE2b-256 d5b4ea0dd91e4bcd9f95d16e135effed69f64f973908bf3ffcf436a3f8693d31

See more details on using hashes here.

File details

Details for the file lookahead_keys_attention-0.0.1-py3-none-any.whl.

File metadata

File hashes

Hashes for lookahead_keys_attention-0.0.1-py3-none-any.whl
Algorithm Hash digest
SHA256 6b181b2ab084ebe39b11a8c43607a9d4814a9c221a2adc5e62c17bc5ddf2e211
MD5 974fe28b133ad4b998917ba722f46a21
BLAKE2b-256 3718cf1a0692f2335e6b090fcceda8b85962ee705130492652e86ce46f352f88

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page