Skip to main content

Lookahead Keys Attention

Project description

Lookahead Keys Attention (wip)

Causal Attention with Lookahead Keys

Installation

pip install lookahead-keys-attention

Usage

import torch
from lookahead_keys_attention import Castle

# Initialize the Castle attention module
model = Castle(
    dim=512,           # input dimension
    heads=8,           # number of attention heads
    dim_head=64,       # dimension per head
    use_triton=None    # auto-detect CUDA for Triton optimization
)

# Example with CUDA sequence
batch_size = 2
seq_len = 128
dim = 512

# Move to CUDA if available
device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')
model = model.to(device)

# Input sequence
x = torch.randn(batch_size, seq_len, dim).to(device)

# Forward pass
output = model(x)  # Shape: [batch_size, seq_len, dim]

# For inference with caching (single token generation)
cache = None
for i in range(seq_len):
    token = x[:, i:i+1, :]  # Single token
    output, cache = model(token, cache=cache, return_next_cache=True)

Citations

@inproceedings{Song2025CausalAW,
    title   = {Causal Attention with Lookahead Keys},
    author  = {Zhuoqing Song and Peng Sun and Huizhuo Yuan and Quanquan Gu},
    year    = {2025},
    url     = {https://api.semanticscholar.org/CorpusID:281218151}
}

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

lookahead_keys_attention-0.0.2.tar.gz (153.2 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

lookahead_keys_attention-0.0.2-py3-none-any.whl (10.3 kB view details)

Uploaded Python 3

File details

Details for the file lookahead_keys_attention-0.0.2.tar.gz.

File metadata

  • Download URL: lookahead_keys_attention-0.0.2.tar.gz
  • Upload date:
  • Size: 153.2 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.2.0 CPython/3.9.23

File hashes

Hashes for lookahead_keys_attention-0.0.2.tar.gz
Algorithm Hash digest
SHA256 5830cf06274f7ae37cb3207e6d183d7bc77d8260400f75edb27c121e50decb1c
MD5 d26be14b046f11fc7dbea6193c03dc29
BLAKE2b-256 84965620f24d30cc2752c710b5214328d44d79e02ddc3400e691566d4c4caabf

See more details on using hashes here.

File details

Details for the file lookahead_keys_attention-0.0.2-py3-none-any.whl.

File metadata

File hashes

Hashes for lookahead_keys_attention-0.0.2-py3-none-any.whl
Algorithm Hash digest
SHA256 01c25b2394a9b5922cecbd8785e9983bf3e06f3d6314b0fa25d6f7e41b9c414f
MD5 9d02151836c8ea80197cc40d65360772
BLAKE2b-256 98e713bde587bbf7766c554d8a3cc353d8104a7679e93737782fc5405cc24083

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page